var/home/core/zuul-output/0000755000175000017500000000000015111064231014517 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111074552015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004755062415111074542017712 0ustar rootrootNov 24 14:19:21 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 14:19:21 crc restorecon[4686]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:21 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 14:19:22 crc restorecon[4686]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 14:19:23 crc kubenswrapper[4822]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 14:19:23 crc kubenswrapper[4822]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 14:19:23 crc kubenswrapper[4822]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 14:19:23 crc kubenswrapper[4822]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 14:19:23 crc kubenswrapper[4822]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 14:19:23 crc kubenswrapper[4822]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.444954 4822 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453093 4822 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453130 4822 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453139 4822 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453150 4822 feature_gate.go:330] unrecognized feature gate: Example Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453162 4822 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453171 4822 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453180 4822 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453189 4822 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453199 4822 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453237 4822 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453246 4822 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453255 4822 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453263 4822 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453286 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453295 4822 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453303 4822 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453311 4822 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453318 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453327 4822 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453337 4822 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453347 4822 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453357 4822 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453368 4822 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453379 4822 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453391 4822 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453402 4822 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453413 4822 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453427 4822 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453441 4822 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453452 4822 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453464 4822 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453474 4822 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453487 4822 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453496 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453504 4822 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453512 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453520 4822 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453528 4822 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453536 4822 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453543 4822 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453556 4822 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453566 4822 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453576 4822 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453585 4822 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453593 4822 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453605 4822 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453614 4822 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453624 4822 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453632 4822 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453640 4822 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453648 4822 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453656 4822 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453664 4822 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453671 4822 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453680 4822 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453687 4822 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453698 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453707 4822 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453714 4822 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453722 4822 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453730 4822 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453737 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453749 4822 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453759 4822 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453767 4822 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453776 4822 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453784 4822 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453792 4822 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453800 4822 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453808 4822 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.453815 4822 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455589 4822 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455633 4822 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455656 4822 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455669 4822 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455682 4822 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455693 4822 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455709 4822 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455721 4822 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455731 4822 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455740 4822 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455751 4822 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455761 4822 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455770 4822 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455779 4822 flags.go:64] FLAG: --cgroup-root="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455788 4822 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455797 4822 flags.go:64] FLAG: --client-ca-file="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455806 4822 flags.go:64] FLAG: --cloud-config="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455816 4822 flags.go:64] FLAG: --cloud-provider="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455825 4822 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455838 4822 flags.go:64] FLAG: --cluster-domain="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455847 4822 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455856 4822 flags.go:64] FLAG: --config-dir="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455866 4822 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455876 4822 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455890 4822 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455902 4822 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455914 4822 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455927 4822 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455941 4822 flags.go:64] FLAG: --contention-profiling="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455953 4822 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455966 4822 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455976 4822 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455985 4822 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.455997 4822 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456007 4822 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456016 4822 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456025 4822 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456034 4822 flags.go:64] FLAG: --enable-server="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456044 4822 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456057 4822 flags.go:64] FLAG: --event-burst="100" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456066 4822 flags.go:64] FLAG: --event-qps="50" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456075 4822 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456086 4822 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456095 4822 flags.go:64] FLAG: --eviction-hard="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456106 4822 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456119 4822 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456128 4822 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456137 4822 flags.go:64] FLAG: --eviction-soft="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456146 4822 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456155 4822 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456163 4822 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456172 4822 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456181 4822 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456189 4822 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456198 4822 flags.go:64] FLAG: --feature-gates="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456245 4822 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456255 4822 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456265 4822 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456275 4822 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456284 4822 flags.go:64] FLAG: --healthz-port="10248" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456294 4822 flags.go:64] FLAG: --help="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456303 4822 flags.go:64] FLAG: --hostname-override="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456312 4822 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456322 4822 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456333 4822 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456345 4822 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456356 4822 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456367 4822 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456378 4822 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456389 4822 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456401 4822 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456412 4822 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456422 4822 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456431 4822 flags.go:64] FLAG: --kube-reserved="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456441 4822 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456449 4822 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456459 4822 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456468 4822 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456479 4822 flags.go:64] FLAG: --lock-file="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456489 4822 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456498 4822 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456507 4822 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456567 4822 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456576 4822 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456585 4822 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456595 4822 flags.go:64] FLAG: --logging-format="text" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456604 4822 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456614 4822 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456624 4822 flags.go:64] FLAG: --manifest-url="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456633 4822 flags.go:64] FLAG: --manifest-url-header="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456646 4822 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456655 4822 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456667 4822 flags.go:64] FLAG: --max-pods="110" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456676 4822 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456685 4822 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456695 4822 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456703 4822 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456713 4822 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456722 4822 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456732 4822 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456758 4822 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456768 4822 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456777 4822 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456787 4822 flags.go:64] FLAG: --pod-cidr="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456797 4822 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456811 4822 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456820 4822 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456829 4822 flags.go:64] FLAG: --pods-per-core="0" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456838 4822 flags.go:64] FLAG: --port="10250" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456848 4822 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456858 4822 flags.go:64] FLAG: --provider-id="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456867 4822 flags.go:64] FLAG: --qos-reserved="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456876 4822 flags.go:64] FLAG: --read-only-port="10255" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456885 4822 flags.go:64] FLAG: --register-node="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456898 4822 flags.go:64] FLAG: --register-schedulable="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456910 4822 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456930 4822 flags.go:64] FLAG: --registry-burst="10" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456942 4822 flags.go:64] FLAG: --registry-qps="5" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456952 4822 flags.go:64] FLAG: --reserved-cpus="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456961 4822 flags.go:64] FLAG: --reserved-memory="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456974 4822 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456985 4822 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.456995 4822 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457004 4822 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457013 4822 flags.go:64] FLAG: --runonce="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457022 4822 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457032 4822 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457041 4822 flags.go:64] FLAG: --seccomp-default="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457050 4822 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457059 4822 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457068 4822 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457079 4822 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457088 4822 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457097 4822 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457106 4822 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457115 4822 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457124 4822 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457134 4822 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457143 4822 flags.go:64] FLAG: --system-cgroups="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457151 4822 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457165 4822 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457175 4822 flags.go:64] FLAG: --tls-cert-file="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457184 4822 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457195 4822 flags.go:64] FLAG: --tls-min-version="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457210 4822 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457247 4822 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457258 4822 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457270 4822 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457281 4822 flags.go:64] FLAG: --v="2" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457297 4822 flags.go:64] FLAG: --version="false" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457314 4822 flags.go:64] FLAG: --vmodule="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457327 4822 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.457340 4822 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457600 4822 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457612 4822 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457620 4822 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457629 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457639 4822 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457647 4822 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457656 4822 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457664 4822 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457672 4822 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457682 4822 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457690 4822 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457699 4822 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457707 4822 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457715 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457723 4822 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457731 4822 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457739 4822 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457746 4822 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457757 4822 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457770 4822 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457779 4822 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457788 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457796 4822 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457806 4822 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457816 4822 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457825 4822 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457834 4822 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457843 4822 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457853 4822 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457862 4822 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457870 4822 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457878 4822 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457887 4822 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457899 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457909 4822 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457920 4822 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457931 4822 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457941 4822 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457951 4822 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457959 4822 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457967 4822 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457975 4822 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457983 4822 feature_gate.go:330] unrecognized feature gate: Example Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.457992 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458000 4822 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458007 4822 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458016 4822 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458024 4822 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458031 4822 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458039 4822 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458050 4822 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458059 4822 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458067 4822 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458075 4822 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458083 4822 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458091 4822 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458099 4822 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458107 4822 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458115 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458123 4822 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458131 4822 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458139 4822 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458146 4822 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458154 4822 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458162 4822 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458170 4822 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458177 4822 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458186 4822 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458193 4822 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458209 4822 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.458245 4822 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.458258 4822 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.470305 4822 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.470351 4822 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470420 4822 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470428 4822 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470433 4822 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470437 4822 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470443 4822 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470453 4822 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470458 4822 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470463 4822 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470468 4822 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470472 4822 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470476 4822 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470479 4822 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470483 4822 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470487 4822 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470491 4822 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470494 4822 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470498 4822 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470501 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470505 4822 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470508 4822 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470511 4822 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470515 4822 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470519 4822 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470523 4822 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470527 4822 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470531 4822 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470535 4822 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470539 4822 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470544 4822 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470548 4822 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470552 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470556 4822 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470561 4822 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470566 4822 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470570 4822 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470574 4822 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470578 4822 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470581 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470585 4822 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470589 4822 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470592 4822 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470596 4822 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470600 4822 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470604 4822 feature_gate.go:330] unrecognized feature gate: Example Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470608 4822 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470612 4822 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470615 4822 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470619 4822 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470622 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470626 4822 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470631 4822 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470635 4822 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470639 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470644 4822 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470648 4822 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470652 4822 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470656 4822 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470660 4822 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470663 4822 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470667 4822 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470670 4822 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470673 4822 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470677 4822 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470680 4822 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470684 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470687 4822 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470691 4822 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470694 4822 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470698 4822 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470701 4822 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.470704 4822 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.470711 4822 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473559 4822 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473618 4822 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473625 4822 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473633 4822 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473639 4822 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473646 4822 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473652 4822 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473658 4822 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473664 4822 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473671 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473679 4822 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473686 4822 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473692 4822 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473701 4822 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473711 4822 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473718 4822 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473724 4822 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473730 4822 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473735 4822 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473741 4822 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473747 4822 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473753 4822 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473759 4822 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473767 4822 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473774 4822 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473781 4822 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473787 4822 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473793 4822 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473800 4822 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473816 4822 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473823 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473829 4822 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473835 4822 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473841 4822 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473847 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473853 4822 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473860 4822 feature_gate.go:330] unrecognized feature gate: Example Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473867 4822 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473875 4822 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473881 4822 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473889 4822 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473896 4822 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473906 4822 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473913 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473920 4822 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473927 4822 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473934 4822 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473940 4822 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473946 4822 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473953 4822 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473959 4822 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473965 4822 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473973 4822 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473979 4822 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473984 4822 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473991 4822 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.473997 4822 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474005 4822 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474012 4822 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474018 4822 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474025 4822 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474032 4822 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474038 4822 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474043 4822 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474048 4822 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474053 4822 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474059 4822 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474064 4822 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474069 4822 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474074 4822 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.474079 4822 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.474090 4822 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.475430 4822 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.482547 4822 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.482747 4822 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.484788 4822 server.go:997] "Starting client certificate rotation" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.484858 4822 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.485152 4822 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-18 04:22:34.507119827 +0000 UTC Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.485290 4822 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.517660 4822 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.520357 4822 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.521131 4822 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.544835 4822 log.go:25] "Validated CRI v1 runtime API" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.577859 4822 log.go:25] "Validated CRI v1 image API" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.580355 4822 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.589085 4822 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-14-14-53-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.589140 4822 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.610150 4822 manager.go:217] Machine: {Timestamp:2025-11-24 14:19:23.605758949 +0000 UTC m=+0.722399476 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:4bde0dbd-3112-410e-b715-484aff4b0fa5 BootID:498687cf-1a90-4e07-8490-79ba0b3ff4d4 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:51:4f:df Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:51:4f:df Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d7:0a:7c Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:95:84:8c Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:23:c3:65 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:5f:f1:28 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:5a:47:a6:61:99:b0 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fa:e7:d5:79:ad:b1 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.610624 4822 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.610845 4822 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.612270 4822 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.612577 4822 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.612622 4822 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.613410 4822 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.613441 4822 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.614120 4822 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.614152 4822 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.614445 4822 state_mem.go:36] "Initialized new in-memory state store" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.614616 4822 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.621141 4822 kubelet.go:418] "Attempting to sync node with API server" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.621193 4822 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.621265 4822 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.621290 4822 kubelet.go:324] "Adding apiserver pod source" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.621310 4822 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.627054 4822 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.628072 4822 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.628650 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.628673 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.628751 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.628778 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.631323 4822 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633208 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633251 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633263 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633273 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633288 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633298 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633307 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633323 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633333 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633345 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633360 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.633370 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.635229 4822 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.635936 4822 server.go:1280] "Started kubelet" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.636627 4822 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.637478 4822 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 14:19:23 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.638685 4822 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.638912 4822 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.641372 4822 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.641415 4822 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.641735 4822 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 15:18:55.540118445 +0000 UTC Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.641810 4822 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1128h59m31.898334814s for next certificate rotation Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.642427 4822 server.go:460] "Adding debug handlers to kubelet server" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.642751 4822 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.642772 4822 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.642860 4822 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.643292 4822 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.643941 4822 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="200ms" Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.643953 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.644058 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.644331 4822 factory.go:55] Registering systemd factory Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.644358 4822 factory.go:221] Registration of the systemd container factory successfully Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.644778 4822 factory.go:153] Registering CRI-O factory Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.644820 4822 factory.go:221] Registration of the crio container factory successfully Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.644955 4822 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.645011 4822 factory.go:103] Registering Raw factory Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.645045 4822 manager.go:1196] Started watching for new ooms in manager Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.644725 4822 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.243:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187af729f472dacc default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 14:19:23.635858124 +0000 UTC m=+0.752498601,LastTimestamp:2025-11-24 14:19:23.635858124 +0000 UTC m=+0.752498601,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.646385 4822 manager.go:319] Starting recovery of all containers Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.661537 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.661680 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.661712 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.661747 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.661775 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.664595 4822 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.664992 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.665316 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.665627 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.665853 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.666473 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.666779 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.666954 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.667120 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.667329 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.667502 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.667692 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.667898 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.667751 4822 manager.go:324] Recovery completed Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.668076 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.668959 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.669121 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.669342 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.669512 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.669666 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.669812 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.669966 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.670127 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.670410 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.670940 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.671249 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.671429 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.671596 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.671841 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.672022 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.672177 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.672401 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.672571 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.672767 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.672920 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.673089 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.673295 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.673465 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.673632 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.673822 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.674004 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.674182 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.674457 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.674650 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.674846 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.675015 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.675169 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.675363 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.675524 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.675736 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.675900 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.676051 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.676272 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.676450 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.676636 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.676819 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.676969 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.677106 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.677324 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.677506 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.677853 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.678029 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.678190 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.678523 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.678702 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.678880 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.679026 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.679253 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.679423 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.679582 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.679767 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.679921 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.680093 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.680293 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.680464 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.680643 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.680842 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.681018 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.681182 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.681393 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.681556 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.681764 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.681930 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.682100 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.681937 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.682943 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.683819 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.684023 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.684147 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.684286 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.684416 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.684519 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.684623 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.684719 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.684836 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.684935 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.685042 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.685149 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.685270 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.685373 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.685464 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.685571 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.685685 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.685793 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.685905 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.686005 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.686106 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.686209 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.686347 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.686451 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.686583 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.686682 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.686785 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.686894 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.687024 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.687121 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.687521 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.687642 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.687772 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.687905 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688019 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688105 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688237 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688361 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688471 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.687794 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688554 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688721 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688833 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688938 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.689041 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.689146 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.689278 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.689389 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.689500 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.689606 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.688618 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.690000 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.689712 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.690301 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.690430 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.690576 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.690700 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.690815 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.690924 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691032 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691112 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691188 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691318 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691408 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691504 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691594 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691705 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691789 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691879 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.691960 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692068 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692171 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692298 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692396 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692475 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692574 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692654 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692730 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692803 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692876 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.692953 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693036 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693124 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693263 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693430 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693521 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693612 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693690 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693764 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693843 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693917 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694020 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694119 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.693198 4822 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694246 4822 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694276 4822 state_mem.go:36] "Initialized new in-memory state store" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694199 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694600 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694701 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694779 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694879 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.694961 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695061 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695159 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695331 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695396 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695412 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695429 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695445 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695460 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695480 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695495 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695510 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695524 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695538 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695556 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695572 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695601 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695616 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695647 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695665 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695689 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695703 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695717 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695735 4822 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695749 4822 reconstruct.go:97] "Volume reconstruction finished" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.695768 4822 reconciler.go:26] "Reconciler: start to sync state" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.700345 4822 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.702284 4822 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.702347 4822 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.703142 4822 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 14:19:23 crc kubenswrapper[4822]: W1124 14:19:23.703290 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.703375 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.703529 4822 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.724271 4822 policy_none.go:49] "None policy: Start" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.725346 4822 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.725386 4822 state_mem.go:35] "Initializing new in-memory state store" Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.743518 4822 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.786403 4822 manager.go:334] "Starting Device Plugin manager" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.786543 4822 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.786556 4822 server.go:79] "Starting device plugin registration server" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.786894 4822 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.786914 4822 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.787116 4822 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.787180 4822 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.787187 4822 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.796871 4822 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.804351 4822 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.804438 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.805697 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.805724 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.805732 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.805847 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.806049 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.806168 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.806426 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.806446 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.806454 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.806922 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.808980 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.809006 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.809023 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.809062 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.809129 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.809163 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.809919 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.810034 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.810101 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.810121 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.810139 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.811165 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.811293 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.811315 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.811648 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.811677 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.811683 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.811728 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.811689 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.811748 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.812791 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.813392 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.813429 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.814303 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.814333 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.814344 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.814343 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.814446 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.814490 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.814757 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.814814 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.815662 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.815689 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.815698 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.845050 4822 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="400ms" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.887664 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.889065 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.889106 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.889119 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.889148 4822 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 14:19:23 crc kubenswrapper[4822]: E1124 14:19:23.889642 4822 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.897957 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898004 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898046 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898122 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898175 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898233 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898263 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898295 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898323 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898353 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898376 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898423 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898462 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898482 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:23 crc kubenswrapper[4822]: I1124 14:19:23.898498 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000019 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000075 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000096 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000115 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000134 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000155 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000172 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000195 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000238 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000258 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000274 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000289 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000273 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000374 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000404 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000330 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000467 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000491 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000504 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000541 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000308 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000572 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000563 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000590 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000546 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000636 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000664 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000706 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000770 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.000860 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.090745 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.092336 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.092392 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.092410 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.092444 4822 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 14:19:24 crc kubenswrapper[4822]: E1124 14:19:24.092958 4822 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.148913 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.157381 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.175120 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.206988 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: W1124 14:19:24.208556 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-57e0611927070156c403c9b3587df6ff2067aeaf780164e8721d59971d37ac66 WatchSource:0}: Error finding container 57e0611927070156c403c9b3587df6ff2067aeaf780164e8721d59971d37ac66: Status 404 returned error can't find the container with id 57e0611927070156c403c9b3587df6ff2067aeaf780164e8721d59971d37ac66 Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.212312 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:19:24 crc kubenswrapper[4822]: W1124 14:19:24.213307 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-f17d2663cfa201166c78a23ea17988f6b997cd3de4d29284e422366b8ee496a3 WatchSource:0}: Error finding container f17d2663cfa201166c78a23ea17988f6b997cd3de4d29284e422366b8ee496a3: Status 404 returned error can't find the container with id f17d2663cfa201166c78a23ea17988f6b997cd3de4d29284e422366b8ee496a3 Nov 24 14:19:24 crc kubenswrapper[4822]: W1124 14:19:24.215097 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-fece62be8d01e4a51ebae7288163cf8133ada4a8fb31c0d7f41304f016a94c4f WatchSource:0}: Error finding container fece62be8d01e4a51ebae7288163cf8133ada4a8fb31c0d7f41304f016a94c4f: Status 404 returned error can't find the container with id fece62be8d01e4a51ebae7288163cf8133ada4a8fb31c0d7f41304f016a94c4f Nov 24 14:19:24 crc kubenswrapper[4822]: W1124 14:19:24.221289 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-cc35eca07a9e28476d4874aa48aee2f8cad7f5d85af2097eeee0d43f807dc559 WatchSource:0}: Error finding container cc35eca07a9e28476d4874aa48aee2f8cad7f5d85af2097eeee0d43f807dc559: Status 404 returned error can't find the container with id cc35eca07a9e28476d4874aa48aee2f8cad7f5d85af2097eeee0d43f807dc559 Nov 24 14:19:24 crc kubenswrapper[4822]: W1124 14:19:24.233860 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c390107f176d3e2326ac238b75aa2c4308d0f0ba729366174755c69343525e6d WatchSource:0}: Error finding container c390107f176d3e2326ac238b75aa2c4308d0f0ba729366174755c69343525e6d: Status 404 returned error can't find the container with id c390107f176d3e2326ac238b75aa2c4308d0f0ba729366174755c69343525e6d Nov 24 14:19:24 crc kubenswrapper[4822]: E1124 14:19:24.246083 4822 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="800ms" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.493579 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.495496 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.495568 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.495584 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.495987 4822 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 14:19:24 crc kubenswrapper[4822]: E1124 14:19:24.498292 4822 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Nov 24 14:19:24 crc kubenswrapper[4822]: W1124 14:19:24.526756 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:24 crc kubenswrapper[4822]: E1124 14:19:24.526847 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.640446 4822 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.710542 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f17d2663cfa201166c78a23ea17988f6b997cd3de4d29284e422366b8ee496a3"} Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.711893 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"57e0611927070156c403c9b3587df6ff2067aeaf780164e8721d59971d37ac66"} Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.713346 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c390107f176d3e2326ac238b75aa2c4308d0f0ba729366174755c69343525e6d"} Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.714665 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cc35eca07a9e28476d4874aa48aee2f8cad7f5d85af2097eeee0d43f807dc559"} Nov 24 14:19:24 crc kubenswrapper[4822]: I1124 14:19:24.715468 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fece62be8d01e4a51ebae7288163cf8133ada4a8fb31c0d7f41304f016a94c4f"} Nov 24 14:19:24 crc kubenswrapper[4822]: W1124 14:19:24.722365 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:24 crc kubenswrapper[4822]: E1124 14:19:24.722456 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:25 crc kubenswrapper[4822]: E1124 14:19:25.047467 4822 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="1.6s" Nov 24 14:19:25 crc kubenswrapper[4822]: W1124 14:19:25.055191 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:25 crc kubenswrapper[4822]: E1124 14:19:25.055494 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:25 crc kubenswrapper[4822]: W1124 14:19:25.089164 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:25 crc kubenswrapper[4822]: E1124 14:19:25.089329 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.299263 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.300752 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.300801 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.300817 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.300850 4822 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 14:19:25 crc kubenswrapper[4822]: E1124 14:19:25.301466 4822 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.566538 4822 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 14:19:25 crc kubenswrapper[4822]: E1124 14:19:25.567839 4822 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.639783 4822 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.722896 4822 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322" exitCode=0 Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.723010 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.723031 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322"} Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.724640 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.724701 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.724724 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.727438 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7"} Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.727481 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318"} Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.727506 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5"} Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.727521 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.727529 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1"} Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.728607 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.728668 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.728691 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.731499 4822 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d" exitCode=0 Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.731598 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d"} Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.731730 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.732947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.732995 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.733012 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.734616 4822 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197" exitCode=0 Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.734740 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.734905 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197"} Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.735797 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.735846 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.735869 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.743235 4822 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877" exitCode=0 Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.743493 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.743488 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877"} Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.744532 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.744565 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.744578 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.745016 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.746397 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.746480 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:25 crc kubenswrapper[4822]: I1124 14:19:25.746564 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.129861 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.640384 4822 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:26 crc kubenswrapper[4822]: E1124 14:19:26.649143 4822 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="3.2s" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.750232 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d"} Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.750317 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de"} Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.750350 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378"} Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.750376 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5"} Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.751527 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0d272d832cbfb5ccdd258a4151ebf2e0324964440edd34bd7e6751f9eff958e5"} Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.751566 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.756625 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.756657 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.756668 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.761458 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.761461 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062"} Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.761493 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb"} Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.761503 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f"} Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.762466 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.762494 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.762506 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.766547 4822 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3" exitCode=0 Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.766646 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.766659 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3"} Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.766644 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.767434 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.767460 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.767469 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.768046 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.768058 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.768067 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.901772 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.903391 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.903433 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.903444 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:26 crc kubenswrapper[4822]: I1124 14:19:26.903476 4822 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 14:19:26 crc kubenswrapper[4822]: E1124 14:19:26.903962 4822 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Nov 24 14:19:27 crc kubenswrapper[4822]: W1124 14:19:27.059491 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:27 crc kubenswrapper[4822]: E1124 14:19:27.059564 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:27 crc kubenswrapper[4822]: W1124 14:19:27.090875 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:27 crc kubenswrapper[4822]: E1124 14:19:27.090971 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:27 crc kubenswrapper[4822]: W1124 14:19:27.210076 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:27 crc kubenswrapper[4822]: E1124 14:19:27.210249 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:27 crc kubenswrapper[4822]: W1124 14:19:27.294622 4822 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Nov 24 14:19:27 crc kubenswrapper[4822]: E1124 14:19:27.294752 4822 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.775217 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c"} Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.775412 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.776431 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.776476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.776492 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.777809 4822 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527" exitCode=0 Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.777953 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.777980 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.778002 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.778400 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.778717 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527"} Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.778856 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.779354 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.779392 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.779409 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.779585 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.779614 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.779630 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.780105 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.780288 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.780415 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.780445 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.780459 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:27 crc kubenswrapper[4822]: I1124 14:19:27.780672 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.794598 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5f60e1b6e031acfddc99eaab21f25b1cf791148cb732891e9c251d2577c064a8"} Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.794671 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.794714 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.794675 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bcdf6e8f860ef2a3419b49f53e2ff550d704467c862b5ce00d3946eb284fb284"} Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.794932 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.794979 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6073ae0db844c158057de597185a6f429e235c6442988a6885b208be5f55560f"} Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.795015 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ca06cc6cb37db7454f12a957471e578968131669b24119e2de9748230dcaa3b9"} Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.795039 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"54579d92e01f8b881c6600bc68e65c8c3afe1a52e6e14250077390ea376fcde5"} Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.796399 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.796446 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.796462 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.796481 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.796553 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:28 crc kubenswrapper[4822]: I1124 14:19:28.796570 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.130698 4822 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.130912 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.525831 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.624386 4822 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.797609 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.797672 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.799721 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.799808 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.799837 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.800340 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.800412 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:29 crc kubenswrapper[4822]: I1124 14:19:29.800436 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.104503 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.106416 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.106489 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.106506 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.106547 4822 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.368477 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.800800 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.800803 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.802583 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.802609 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.802646 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.802660 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.802687 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:30 crc kubenswrapper[4822]: I1124 14:19:30.802667 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.313489 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.313717 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.315456 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.315529 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.315550 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.322799 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.803317 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.806760 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.808071 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:31 crc kubenswrapper[4822]: I1124 14:19:31.808134 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.066518 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.066911 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.068788 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.068851 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.068869 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.212250 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.463915 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.805675 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.808253 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.808324 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:32 crc kubenswrapper[4822]: I1124 14:19:32.808351 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.375403 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.375759 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.377581 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.377646 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.377670 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.626220 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.626447 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.627927 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.628011 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.628029 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:33 crc kubenswrapper[4822]: E1124 14:19:33.797258 4822 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.808138 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.809541 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.809602 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:33 crc kubenswrapper[4822]: I1124 14:19:33.809621 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.640664 4822 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.702426 4822 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37176->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.702526 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37176->192.168.126.11:17697: read: connection reset by peer" Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.823092 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.826303 4822 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c" exitCode=255 Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.826353 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c"} Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.826570 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.827867 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.827894 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.827904 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:37 crc kubenswrapper[4822]: I1124 14:19:37.828341 4822 scope.go:117] "RemoveContainer" containerID="88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c" Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.052378 4822 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.052473 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.062292 4822 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.062373 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.832868 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.836259 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b"} Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.836485 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.838009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.838080 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:38 crc kubenswrapper[4822]: I1124 14:19:38.838102 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:39 crc kubenswrapper[4822]: I1124 14:19:39.131247 4822 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 14:19:39 crc kubenswrapper[4822]: I1124 14:19:39.131337 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.396234 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.396498 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.398240 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.398302 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.398323 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.415185 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.842033 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.843491 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.843538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:40 crc kubenswrapper[4822]: I1124 14:19:40.843552 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.074741 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.074968 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.075029 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.076888 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.077099 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.077318 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.081869 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.471359 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.471614 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.473188 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.473263 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.473280 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.848490 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.849835 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.849897 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:42 crc kubenswrapper[4822]: I1124 14:19:42.849939 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.053458 4822 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.054873 4822 trace.go:236] Trace[1099266905]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 14:19:31.419) (total time: 11635ms): Nov 24 14:19:43 crc kubenswrapper[4822]: Trace[1099266905]: ---"Objects listed" error: 11635ms (14:19:43.054) Nov 24 14:19:43 crc kubenswrapper[4822]: Trace[1099266905]: [11.635151191s] [11.635151191s] END Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.054921 4822 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.055942 4822 trace.go:236] Trace[1445246862]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 14:19:31.415) (total time: 11640ms): Nov 24 14:19:43 crc kubenswrapper[4822]: Trace[1445246862]: ---"Objects listed" error: 11640ms (14:19:43.055) Nov 24 14:19:43 crc kubenswrapper[4822]: Trace[1445246862]: [11.640132128s] [11.640132128s] END Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.055972 4822 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.061274 4822 trace.go:236] Trace[799923863]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 14:19:31.376) (total time: 11684ms): Nov 24 14:19:43 crc kubenswrapper[4822]: Trace[799923863]: ---"Objects listed" error: 11684ms (14:19:43.060) Nov 24 14:19:43 crc kubenswrapper[4822]: Trace[799923863]: [11.684255618s] [11.684255618s] END Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.061320 4822 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.063165 4822 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.063939 4822 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.067286 4822 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.068344 4822 trace.go:236] Trace[825108879]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 14:19:32.684) (total time: 10383ms): Nov 24 14:19:43 crc kubenswrapper[4822]: Trace[825108879]: ---"Objects listed" error: 10383ms (14:19:43.068) Nov 24 14:19:43 crc kubenswrapper[4822]: Trace[825108879]: [10.383456509s] [10.383456509s] END Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.068390 4822 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.632030 4822 apiserver.go:52] "Watching apiserver" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.636579 4822 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.637058 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.637689 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.637850 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.637880 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.638007 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.637885 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.638250 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.638482 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.638626 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.638743 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.640051 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.641276 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.642651 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.643010 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.643201 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.643509 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.643567 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.643635 4822 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.643637 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.643943 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.667665 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.667756 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.667814 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.667864 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.667909 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.667954 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668002 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668046 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668090 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668135 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668179 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668260 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668314 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668365 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668414 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668423 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668465 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668489 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668510 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668604 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668666 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668796 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668809 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668851 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668901 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668946 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.668992 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669112 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669122 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669121 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669172 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669179 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669248 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669294 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669341 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669345 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669379 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669383 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669449 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669472 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669491 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669511 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669545 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669563 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669582 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669584 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669603 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669690 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669742 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669794 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669843 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669935 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669985 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670030 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670079 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670125 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670174 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670255 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670303 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670354 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670404 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670453 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670500 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670597 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670640 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670715 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670768 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670815 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670864 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670906 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670951 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.671001 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.671046 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.671094 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.671142 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.671189 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.676287 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.676351 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.676400 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.676451 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669745 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669782 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669841 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669968 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669964 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.669917 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670110 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670126 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670304 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670413 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670431 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.670573 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.671184 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.672598 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.672547 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.672650 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.672892 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.672909 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.673166 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.673392 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.673535 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.673821 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.673849 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.674013 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.674270 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.674427 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.674561 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.674747 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.675233 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.675251 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.677618 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.677698 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.677940 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.678291 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.678506 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.678523 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.678673 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.678665 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.678720 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.677705 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.679446 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.679795 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.680045 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.680147 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.680134 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.680638 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681137 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681274 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.676497 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681334 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681346 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681432 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681523 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681559 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681593 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681628 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681657 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681660 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681697 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681721 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681739 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681757 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681781 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681804 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681828 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681856 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.681901 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.682026 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.682066 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.682311 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.682531 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.682563 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.682583 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.682633 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.682655 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.682865 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683010 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683067 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683067 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683106 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683226 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683251 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683285 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683351 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683373 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683391 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683410 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683709 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683734 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683752 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683770 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683785 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683801 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683819 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683837 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683852 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683868 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683883 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683897 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683915 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683933 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683949 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683964 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683982 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683998 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684014 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684029 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684046 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684063 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684081 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684096 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684111 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684126 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684143 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684158 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684174 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684192 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684223 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684250 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684267 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684285 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684302 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684318 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684336 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684351 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684368 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684383 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684397 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684412 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684428 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684513 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684558 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684578 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684592 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684606 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684621 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684635 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684649 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684663 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684680 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684697 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684713 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684729 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684745 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684762 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684784 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684800 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684815 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684831 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684847 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684864 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684880 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684898 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684914 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684928 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684944 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684960 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684974 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684989 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685005 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685020 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685035 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685051 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685066 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685082 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685097 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685113 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685129 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685144 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685167 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685182 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685218 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685244 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685267 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685289 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685311 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685378 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685403 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685426 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685476 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685509 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685652 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685922 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685957 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685989 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686011 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686046 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686070 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686090 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686107 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686123 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686142 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686158 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686197 4822 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686247 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686262 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686272 4822 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686282 4822 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686292 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686302 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686311 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686320 4822 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686328 4822 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686339 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686349 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686358 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686367 4822 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686376 4822 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686384 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686395 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686404 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686414 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686423 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686431 4822 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686441 4822 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686449 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686459 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686467 4822 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686476 4822 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686484 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686493 4822 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686517 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686527 4822 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686535 4822 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686545 4822 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686554 4822 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686563 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686576 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686585 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686595 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686603 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686612 4822 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686621 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686630 4822 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686642 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686653 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686662 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686671 4822 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686679 4822 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686689 4822 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686698 4822 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686707 4822 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686716 4822 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686727 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686737 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686747 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686757 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686766 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686775 4822 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686785 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686794 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686802 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686812 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686820 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686829 4822 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686837 4822 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686847 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.687556 4822 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683143 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.683279 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.684681 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685384 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.685458 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686325 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.686600 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.687774 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.688483 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.688583 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.688856 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.688885 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.689798 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.690603 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.691012 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.691854 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.692315 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.692711 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.693065 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.694326 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.694692 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.694952 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.695200 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.695800 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.696066 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.696741 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.696879 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.697135 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.697525 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.698538 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.699280 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.699311 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.699683 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.699964 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.700294 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.700889 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.701264 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.701733 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.701871 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.702179 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.702412 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.702511 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.702742 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.702958 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.702958 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.703247 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.707546 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.707733 4822 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.707832 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:44.207800682 +0000 UTC m=+21.324441279 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.708413 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.709578 4822 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.709784 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:44.209746863 +0000 UTC m=+21.326387380 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.711297 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.712084 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.710240 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.713405 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.713700 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.713945 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.714688 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.715708 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.715801 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.715849 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.716124 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.716670 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.717657 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.717991 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.718234 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.718368 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.719078 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.719261 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.719526 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.719624 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.719852 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.720165 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.720453 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.720665 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.721053 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.721485 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.721475 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.721606 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:19:44.221578113 +0000 UTC m=+21.338218690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.721776 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.721823 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.722298 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.722415 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.722537 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.722606 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.722617 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.723048 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.723048 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.723061 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.723446 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.723820 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.723886 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.724047 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.724341 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.724493 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.724647 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.724712 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.724828 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.725377 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.725779 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.726068 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.728663 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.733765 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.733806 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.733822 4822 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.733903 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:44.233881169 +0000 UTC m=+21.350521656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.738363 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.739139 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.739327 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.739686 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.739720 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.739734 4822 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:43 crc kubenswrapper[4822]: E1124 14:19:43.739793 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:44.239773983 +0000 UTC m=+21.356414460 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.739060 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.740312 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.743252 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.743274 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.743591 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.744375 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.746923 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.747697 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.751755 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.752410 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.752616 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.753529 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.754477 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.754615 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.754831 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.755627 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.764718 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.764889 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.765395 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.765539 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.765725 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.765983 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.775046 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.781468 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.782550 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.783261 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.786472 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790294 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790336 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790391 4822 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790401 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790410 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790419 4822 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790427 4822 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790434 4822 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790442 4822 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790450 4822 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790459 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790469 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790477 4822 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790485 4822 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790493 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790501 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790509 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790517 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790525 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790532 4822 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790541 4822 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790549 4822 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790556 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790564 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790573 4822 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790581 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790590 4822 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790598 4822 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790606 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790614 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790622 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790630 4822 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790637 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790645 4822 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790654 4822 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790661 4822 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790670 4822 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790678 4822 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790685 4822 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790693 4822 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790701 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790709 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790717 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790725 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790733 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790741 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790749 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790758 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790765 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790773 4822 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790781 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790788 4822 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790796 4822 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790806 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790814 4822 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790821 4822 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790830 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790838 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790849 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790857 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790865 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790874 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790882 4822 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790890 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790897 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790905 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790912 4822 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790921 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790929 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790937 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790944 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790951 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790959 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790966 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790975 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790982 4822 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790990 4822 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.790998 4822 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791005 4822 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791012 4822 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791020 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791027 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791036 4822 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791044 4822 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791053 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791061 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791069 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791077 4822 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791085 4822 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791092 4822 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791100 4822 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791108 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791116 4822 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791125 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791132 4822 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791141 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791150 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791157 4822 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791165 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791173 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791180 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791188 4822 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791195 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791216 4822 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791224 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791231 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791239 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791247 4822 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791254 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791262 4822 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791269 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791277 4822 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791284 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791292 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791300 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791308 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791317 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791325 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791333 4822 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791340 4822 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791319 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791623 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791793 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791824 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.791878 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.792108 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.792184 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.792851 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.793795 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.794558 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.795158 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.796309 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.799435 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.799523 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.800493 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.802683 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.804466 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.805445 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.807311 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.808663 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.816478 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.818627 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.819106 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.819789 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.821425 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.822197 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.822380 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.823163 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.826896 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.827655 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.828778 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.831655 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.832970 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.835632 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.836271 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.836604 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.836777 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.837138 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.837141 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.838063 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.838511 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.839882 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.840462 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.840548 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.841386 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.841910 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.842866 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.843410 4822 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.843509 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.845127 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.846035 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.846632 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.848176 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.848302 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.849309 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.849803 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.851061 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.851746 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.852627 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.853238 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.854258 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.854823 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.855641 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.856142 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.857106 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.857885 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.858775 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.859201 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.860047 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.860679 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.861222 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.862351 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.863104 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.874173 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.880362 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.889752 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892135 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892166 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892179 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892190 4822 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892222 4822 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892234 4822 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892245 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892256 4822 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892267 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892278 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892289 4822 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.892301 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.898854 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.910800 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.920480 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.929523 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.961895 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 14:19:43 crc kubenswrapper[4822]: I1124 14:19:43.979875 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.011383 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 14:19:44 crc kubenswrapper[4822]: W1124 14:19:44.024772 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-e87db7529dde123b28050f7879e9473fac1fbeb55334cec712cbc31e81be27dc WatchSource:0}: Error finding container e87db7529dde123b28050f7879e9473fac1fbeb55334cec712cbc31e81be27dc: Status 404 returned error can't find the container with id e87db7529dde123b28050f7879e9473fac1fbeb55334cec712cbc31e81be27dc Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.295976 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.296074 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.296104 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.296125 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.296147 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296266 4822 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296296 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:19:45.296184181 +0000 UTC m=+22.412824668 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296318 4822 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296339 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:45.296329256 +0000 UTC m=+22.412969743 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296338 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296367 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296384 4822 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296367 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:45.296351406 +0000 UTC m=+22.412991883 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296443 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:45.296428059 +0000 UTC m=+22.413068546 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296522 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296565 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296581 4822 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:44 crc kubenswrapper[4822]: E1124 14:19:44.296649 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:45.296610475 +0000 UTC m=+22.413250962 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.855003 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564"} Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.855052 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e050a2b61d303d67c334d2c4de10b7832989414f508295c45fcc5512c302c4a8"} Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.856710 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e87db7529dde123b28050f7879e9473fac1fbeb55334cec712cbc31e81be27dc"} Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.858327 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe"} Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.858384 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6"} Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.858398 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9963e8bc0dee037641d0f1f9efb8a5a433b42cfdf0881eaa15c7f80d20e61c17"} Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.871879 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.888050 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.906736 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.921764 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.939018 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:44Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.952414 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:44Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.971139 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:44Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:44 crc kubenswrapper[4822]: I1124 14:19:44.986948 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:44Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.002750 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:45Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.020146 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:45Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.036602 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:45Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.055945 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:45Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.076729 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:45Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.098371 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:45Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.305908 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306113 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:19:47.306082124 +0000 UTC m=+24.422722601 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.306241 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.306326 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.306393 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.306494 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306437 4822 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306617 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306636 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306651 4822 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306670 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:47.306645532 +0000 UTC m=+24.423286029 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306493 4822 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306562 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306708 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306715 4822 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306697 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:47.306684743 +0000 UTC m=+24.423325300 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306747 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:47.306737825 +0000 UTC m=+24.423378302 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.306759 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:47.306753375 +0000 UTC m=+24.423393912 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.704293 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.704298 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.704470 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.704337 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.704618 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:19:45 crc kubenswrapper[4822]: E1124 14:19:45.704686 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:19:45 crc kubenswrapper[4822]: I1124 14:19:45.710013 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.137629 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.144565 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.152768 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.160998 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.179608 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.196168 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.214844 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.239460 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.254225 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.268990 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.290325 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.307418 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.319366 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.331508 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.344194 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.357817 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.371468 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.385046 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.533619 4822 csr.go:261] certificate signing request csr-wj66l is approved, waiting to be issued Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.576330 4822 csr.go:257] certificate signing request csr-wj66l is issued Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.866760 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4"} Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.882712 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.897419 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.915476 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.927415 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.940064 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.963711 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.978177 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:46 crc kubenswrapper[4822]: I1124 14:19:46.991896 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.322748 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.322824 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.322847 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.322868 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.322890 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.322912 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:19:51.322888757 +0000 UTC m=+28.439529234 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.322997 4822 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323055 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323094 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323011 4822 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323109 4822 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323016 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323093 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:51.323071143 +0000 UTC m=+28.439711620 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323261 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:51.323239269 +0000 UTC m=+28.439879756 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323222 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323300 4822 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323283 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:51.32327298 +0000 UTC m=+28.439913567 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.323338 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:51.323329961 +0000 UTC m=+28.439970428 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.429095 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-kng5k"] Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.429561 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-d6rbt"] Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.429911 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.430306 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-kng5k" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.433177 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.433266 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.433265 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-nqrzx"] Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.433694 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.433825 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.434004 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-nst99"] Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.434166 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.434324 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.439055 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.439336 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.439635 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.440013 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.440805 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.440956 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.441104 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.441509 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.441649 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.441798 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.446555 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.456587 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.471694 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.488049 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.501979 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524362 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524392 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-socket-dir-parent\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524554 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-proxy-tls\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524577 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-run-multus-certs\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524604 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8dsg\" (UniqueName: \"kubernetes.io/projected/45082ac5-a419-4ce2-a11b-e9a6b8178698-kube-api-access-q8dsg\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524629 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-var-lib-cni-multus\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524644 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8qtc\" (UniqueName: \"kubernetes.io/projected/92c230aa-aed2-4b07-82bd-98b1d2ffa456-kube-api-access-f8qtc\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524678 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3f878d17-84e9-40c4-b161-eab718c34373-hosts-file\") pod \"node-resolver-kng5k\" (UID: \"3f878d17-84e9-40c4-b161-eab718c34373\") " pod="openshift-dns/node-resolver-kng5k" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524731 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-cni-dir\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524785 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-os-release\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524816 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524866 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-system-cni-dir\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524900 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-cnibin\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524928 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-system-cni-dir\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524964 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-os-release\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524982 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-hostroot\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.524997 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-etc-kubernetes\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525017 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-cnibin\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525048 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-rootfs\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525065 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkh7k\" (UniqueName: \"kubernetes.io/projected/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-kube-api-access-nkh7k\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525101 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-daemon-config\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525115 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-var-lib-kubelet\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525131 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-mcd-auth-proxy-config\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525159 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/92c230aa-aed2-4b07-82bd-98b1d2ffa456-cni-binary-copy\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525232 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-run-netns\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525269 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4l2l\" (UniqueName: \"kubernetes.io/projected/3f878d17-84e9-40c4-b161-eab718c34373-kube-api-access-f4l2l\") pod \"node-resolver-kng5k\" (UID: \"3f878d17-84e9-40c4-b161-eab718c34373\") " pod="openshift-dns/node-resolver-kng5k" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525291 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-run-k8s-cni-cncf-io\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525320 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-conf-dir\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525338 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/45082ac5-a419-4ce2-a11b-e9a6b8178698-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525389 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-var-lib-cni-bin\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.525431 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/45082ac5-a419-4ce2-a11b-e9a6b8178698-cni-binary-copy\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.537501 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.550296 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.562795 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.578111 4822 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-24 14:14:46 +0000 UTC, rotation deadline is 2026-10-04 06:07:32.350547066 +0000 UTC Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.578163 4822 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7527h47m44.772387049s for next certificate rotation Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.579163 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.590892 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.605509 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.623004 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.625970 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-conf-dir\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626005 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/45082ac5-a419-4ce2-a11b-e9a6b8178698-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626027 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4l2l\" (UniqueName: \"kubernetes.io/projected/3f878d17-84e9-40c4-b161-eab718c34373-kube-api-access-f4l2l\") pod \"node-resolver-kng5k\" (UID: \"3f878d17-84e9-40c4-b161-eab718c34373\") " pod="openshift-dns/node-resolver-kng5k" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626051 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-run-k8s-cni-cncf-io\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626074 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/45082ac5-a419-4ce2-a11b-e9a6b8178698-cni-binary-copy\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626126 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-conf-dir\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626163 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-var-lib-cni-bin\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626184 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-socket-dir-parent\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626242 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-proxy-tls\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626262 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-run-multus-certs\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626297 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8dsg\" (UniqueName: \"kubernetes.io/projected/45082ac5-a419-4ce2-a11b-e9a6b8178698-kube-api-access-q8dsg\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626331 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-var-lib-cni-multus\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626350 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8qtc\" (UniqueName: \"kubernetes.io/projected/92c230aa-aed2-4b07-82bd-98b1d2ffa456-kube-api-access-f8qtc\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626369 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-cni-dir\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626386 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-os-release\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626407 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626429 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3f878d17-84e9-40c4-b161-eab718c34373-hosts-file\") pod \"node-resolver-kng5k\" (UID: \"3f878d17-84e9-40c4-b161-eab718c34373\") " pod="openshift-dns/node-resolver-kng5k" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626451 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-cnibin\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626477 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-system-cni-dir\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626496 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-os-release\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626516 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-system-cni-dir\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626533 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-hostroot\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626558 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-etc-kubernetes\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626578 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-cnibin\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626629 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-rootfs\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626648 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkh7k\" (UniqueName: \"kubernetes.io/projected/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-kube-api-access-nkh7k\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626666 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-daemon-config\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626696 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-var-lib-kubelet\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626715 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-mcd-auth-proxy-config\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626735 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/92c230aa-aed2-4b07-82bd-98b1d2ffa456-cni-binary-copy\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626755 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-run-netns\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626873 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-run-netns\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.626243 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-run-k8s-cni-cncf-io\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627046 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-var-lib-cni-bin\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627251 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/45082ac5-a419-4ce2-a11b-e9a6b8178698-cni-binary-copy\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627263 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-cni-dir\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627334 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/45082ac5-a419-4ce2-a11b-e9a6b8178698-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627409 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-cnibin\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627496 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3f878d17-84e9-40c4-b161-eab718c34373-hosts-file\") pod \"node-resolver-kng5k\" (UID: \"3f878d17-84e9-40c4-b161-eab718c34373\") " pod="openshift-dns/node-resolver-kng5k" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627559 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-rootfs\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627591 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-run-multus-certs\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627647 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-socket-dir-parent\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627651 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-cnibin\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627722 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-var-lib-cni-multus\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627814 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-system-cni-dir\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.628398 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.628450 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/92c230aa-aed2-4b07-82bd-98b1d2ffa456-multus-daemon-config\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627886 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/45082ac5-a419-4ce2-a11b-e9a6b8178698-os-release\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.628038 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-host-var-lib-kubelet\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.628057 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-etc-kubernetes\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627811 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-os-release\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.628524 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/92c230aa-aed2-4b07-82bd-98b1d2ffa456-cni-binary-copy\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.627880 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-hostroot\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.628550 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92c230aa-aed2-4b07-82bd-98b1d2ffa456-system-cni-dir\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.628508 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-mcd-auth-proxy-config\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.632057 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-proxy-tls\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.643057 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8dsg\" (UniqueName: \"kubernetes.io/projected/45082ac5-a419-4ce2-a11b-e9a6b8178698-kube-api-access-q8dsg\") pod \"multus-additional-cni-plugins-nqrzx\" (UID: \"45082ac5-a419-4ce2-a11b-e9a6b8178698\") " pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.644925 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkh7k\" (UniqueName: \"kubernetes.io/projected/9cb8d340-5782-4aa4-b81e-e1b64e6b5207-kube-api-access-nkh7k\") pod \"machine-config-daemon-nst99\" (UID: \"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\") " pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.646830 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.647389 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8qtc\" (UniqueName: \"kubernetes.io/projected/92c230aa-aed2-4b07-82bd-98b1d2ffa456-kube-api-access-f8qtc\") pod \"multus-d6rbt\" (UID: \"92c230aa-aed2-4b07-82bd-98b1d2ffa456\") " pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.650060 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4l2l\" (UniqueName: \"kubernetes.io/projected/3f878d17-84e9-40c4-b161-eab718c34373-kube-api-access-f4l2l\") pod \"node-resolver-kng5k\" (UID: \"3f878d17-84e9-40c4-b161-eab718c34373\") " pod="openshift-dns/node-resolver-kng5k" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.662195 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.674040 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.688187 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.702950 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.704154 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.704156 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.704314 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.704777 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.704854 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:47 crc kubenswrapper[4822]: E1124 14:19:47.705049 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.718382 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.735226 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.749378 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-d6rbt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.755265 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.759433 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-kng5k" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.768518 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.777358 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:19:47 crc kubenswrapper[4822]: W1124 14:19:47.782050 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f878d17_84e9_40c4_b161_eab718c34373.slice/crio-ca755dde4b00cf850b325e113b3dbd52dc5fe021d7341db53f97c8b77cbdad7c WatchSource:0}: Error finding container ca755dde4b00cf850b325e113b3dbd52dc5fe021d7341db53f97c8b77cbdad7c: Status 404 returned error can't find the container with id ca755dde4b00cf850b325e113b3dbd52dc5fe021d7341db53f97c8b77cbdad7c Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.788193 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: W1124 14:19:47.808989 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cb8d340_5782_4aa4_b81e_e1b64e6b5207.slice/crio-b0a2bf1085c609f14e416f50dd3d2cb5f0c045f79f3e2506ed0c0b2bfca516ec WatchSource:0}: Error finding container b0a2bf1085c609f14e416f50dd3d2cb5f0c045f79f3e2506ed0c0b2bfca516ec: Status 404 returned error can't find the container with id b0a2bf1085c609f14e416f50dd3d2cb5f0c045f79f3e2506ed0c0b2bfca516ec Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.809088 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.833400 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-swpcw"] Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.834553 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.837673 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.838322 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.838456 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.838567 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.838677 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.838784 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.839263 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.849671 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.862330 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.872125 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.872357 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerStarted","Data":"16f56ac6b22a3dafce01ed30fac547675246c85b0e68b760803130f35678adc7"} Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.873540 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-kng5k" event={"ID":"3f878d17-84e9-40c4-b161-eab718c34373","Type":"ContainerStarted","Data":"ca755dde4b00cf850b325e113b3dbd52dc5fe021d7341db53f97c8b77cbdad7c"} Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.874993 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-d6rbt" event={"ID":"92c230aa-aed2-4b07-82bd-98b1d2ffa456","Type":"ContainerStarted","Data":"843e329611629595117b8b32ec1b219947e31bc538bf76c0b58f60579d30a939"} Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.876195 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"b0a2bf1085c609f14e416f50dd3d2cb5f0c045f79f3e2506ed0c0b2bfca516ec"} Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.887713 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.903259 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.914751 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929420 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-ovn\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929474 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-etc-openvswitch\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929496 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-log-socket\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929525 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-script-lib\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929554 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-slash\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929569 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-netd\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929595 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-bin\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929622 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-systemd\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929639 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-systemd-units\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929656 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-netns\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929672 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-ovn-kubernetes\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929691 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929711 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovn-node-metrics-cert\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929726 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-env-overrides\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929742 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-kubelet\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929758 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhtnb\" (UniqueName: \"kubernetes.io/projected/13d10aa8-fab6-40df-90c8-0784ebac8e0f-kube-api-access-nhtnb\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929775 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-openvswitch\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929789 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-node-log\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929807 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-var-lib-openvswitch\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.929821 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-config\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.955187 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:47 crc kubenswrapper[4822]: I1124 14:19:47.977422 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.002717 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.017171 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030090 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-slash\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030125 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-netd\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030153 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-systemd\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030169 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-bin\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030184 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-systemd-units\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030215 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-netns\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030231 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-ovn-kubernetes\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030248 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030268 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovn-node-metrics-cert\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030254 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-netd\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030286 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-kubelet\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030285 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-slash\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030328 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-env-overrides\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030339 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-systemd\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030340 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-kubelet\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030354 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhtnb\" (UniqueName: \"kubernetes.io/projected/13d10aa8-fab6-40df-90c8-0784ebac8e0f-kube-api-access-nhtnb\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030374 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-netns\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030380 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-ovn-kubernetes\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030377 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-openvswitch\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030396 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-openvswitch\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030409 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030417 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-node-log\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030438 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-var-lib-openvswitch\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030588 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-config\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030607 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-ovn\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030651 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-etc-openvswitch\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030667 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-log-socket\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030708 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-script-lib\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.030909 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-env-overrides\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.031083 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-var-lib-openvswitch\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.031130 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-bin\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.031147 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-ovn\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.031159 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-systemd-units\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.031183 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-node-log\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.031224 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-etc-openvswitch\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.031248 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-log-socket\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.031292 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-script-lib\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.031597 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-config\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.033599 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovn-node-metrics-cert\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.039625 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.046266 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhtnb\" (UniqueName: \"kubernetes.io/projected/13d10aa8-fab6-40df-90c8-0784ebac8e0f-kube-api-access-nhtnb\") pod \"ovnkube-node-swpcw\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.052124 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.064244 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.156926 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:48 crc kubenswrapper[4822]: W1124 14:19:48.167961 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13d10aa8_fab6_40df_90c8_0784ebac8e0f.slice/crio-c26d06b1492bc2eb2fff84b5cc56e70eeb295186b1a80d88755064083c67754a WatchSource:0}: Error finding container c26d06b1492bc2eb2fff84b5cc56e70eeb295186b1a80d88755064083c67754a: Status 404 returned error can't find the container with id c26d06b1492bc2eb2fff84b5cc56e70eeb295186b1a80d88755064083c67754a Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.880519 4822 generic.go:334] "Generic (PLEG): container finished" podID="45082ac5-a419-4ce2-a11b-e9a6b8178698" containerID="dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae" exitCode=0 Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.880597 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerDied","Data":"dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae"} Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.882673 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-kng5k" event={"ID":"3f878d17-84e9-40c4-b161-eab718c34373","Type":"ContainerStarted","Data":"c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09"} Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.884592 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-d6rbt" event={"ID":"92c230aa-aed2-4b07-82bd-98b1d2ffa456","Type":"ContainerStarted","Data":"385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16"} Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.886130 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7" exitCode=0 Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.886177 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7"} Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.886228 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"c26d06b1492bc2eb2fff84b5cc56e70eeb295186b1a80d88755064083c67754a"} Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.888417 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd"} Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.888449 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f"} Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.896448 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.912943 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.929476 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.950229 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.965608 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.983581 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:48 crc kubenswrapper[4822]: I1124 14:19:48.998080 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.012533 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.027191 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.039420 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.059395 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.081696 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.119705 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.132672 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.151566 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.171796 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.184782 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.199262 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.211574 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.223542 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.235950 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.250748 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.261977 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.273235 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.285992 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.299894 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.463508 4822 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.465755 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.465802 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.465817 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.465904 4822 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.472454 4822 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.472737 4822 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.474288 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.474321 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.474333 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.474351 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.474372 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:49 crc kubenswrapper[4822]: E1124 14:19:49.498379 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.502189 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.502258 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.502271 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.502288 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.502300 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:49 crc kubenswrapper[4822]: E1124 14:19:49.517542 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.521010 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.521062 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.521074 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.521091 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.521456 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:49 crc kubenswrapper[4822]: E1124 14:19:49.537818 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.541137 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.541170 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.541179 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.541194 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.541226 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:49 crc kubenswrapper[4822]: E1124 14:19:49.556961 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.562111 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.562179 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.562196 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.562243 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.562260 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:49 crc kubenswrapper[4822]: E1124 14:19:49.580663 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: E1124 14:19:49.581105 4822 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.583217 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.583251 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.583270 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.583288 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.583299 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.686803 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.686838 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.686847 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.686862 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.686871 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.703486 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.703531 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.703492 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:49 crc kubenswrapper[4822]: E1124 14:19:49.703654 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:19:49 crc kubenswrapper[4822]: E1124 14:19:49.703770 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:19:49 crc kubenswrapper[4822]: E1124 14:19:49.703905 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.790084 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.790482 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.790497 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.790515 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.790528 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.892383 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.892878 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.892894 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.892912 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.892922 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.893935 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerStarted","Data":"b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.896737 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.896761 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.896771 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.896780 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.896788 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766"} Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.912085 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.925880 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.942626 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.957513 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.968502 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.985318 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:49Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.995121 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.995151 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.995160 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.995174 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:49 crc kubenswrapper[4822]: I1124 14:19:49.995183 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:49Z","lastTransitionTime":"2025-11-24T14:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.003463 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.015319 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.027890 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.039869 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.051621 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.061763 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.072414 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.083078 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-x8rxd"] Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.083501 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.085285 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.085285 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.085633 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.085644 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.097921 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.097947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.097956 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.097971 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.097981 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:50Z","lastTransitionTime":"2025-11-24T14:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.099972 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.112811 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.122621 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.135349 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.150792 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7ckq\" (UniqueName: \"kubernetes.io/projected/d05c7955-ab58-4997-847d-559aa181ea8e-kube-api-access-r7ckq\") pod \"node-ca-x8rxd\" (UID: \"d05c7955-ab58-4997-847d-559aa181ea8e\") " pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.151045 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d05c7955-ab58-4997-847d-559aa181ea8e-host\") pod \"node-ca-x8rxd\" (UID: \"d05c7955-ab58-4997-847d-559aa181ea8e\") " pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.151157 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d05c7955-ab58-4997-847d-559aa181ea8e-serviceca\") pod \"node-ca-x8rxd\" (UID: \"d05c7955-ab58-4997-847d-559aa181ea8e\") " pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.151021 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.173857 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.200578 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.201042 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.201128 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.201183 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.201278 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.201335 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:50Z","lastTransitionTime":"2025-11-24T14:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.220519 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.231742 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.244259 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.252174 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7ckq\" (UniqueName: \"kubernetes.io/projected/d05c7955-ab58-4997-847d-559aa181ea8e-kube-api-access-r7ckq\") pod \"node-ca-x8rxd\" (UID: \"d05c7955-ab58-4997-847d-559aa181ea8e\") " pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.252233 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d05c7955-ab58-4997-847d-559aa181ea8e-host\") pod \"node-ca-x8rxd\" (UID: \"d05c7955-ab58-4997-847d-559aa181ea8e\") " pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.252254 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d05c7955-ab58-4997-847d-559aa181ea8e-serviceca\") pod \"node-ca-x8rxd\" (UID: \"d05c7955-ab58-4997-847d-559aa181ea8e\") " pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.252564 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d05c7955-ab58-4997-847d-559aa181ea8e-host\") pod \"node-ca-x8rxd\" (UID: \"d05c7955-ab58-4997-847d-559aa181ea8e\") " pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.253340 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d05c7955-ab58-4997-847d-559aa181ea8e-serviceca\") pod \"node-ca-x8rxd\" (UID: \"d05c7955-ab58-4997-847d-559aa181ea8e\") " pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.257772 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.271708 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.273819 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7ckq\" (UniqueName: \"kubernetes.io/projected/d05c7955-ab58-4997-847d-559aa181ea8e-kube-api-access-r7ckq\") pod \"node-ca-x8rxd\" (UID: \"d05c7955-ab58-4997-847d-559aa181ea8e\") " pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.288441 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.302609 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.303970 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.304003 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.304015 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.304034 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.304046 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:50Z","lastTransitionTime":"2025-11-24T14:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.400724 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-x8rxd" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.410867 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.410902 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.410912 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.410925 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.410935 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:50Z","lastTransitionTime":"2025-11-24T14:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: W1124 14:19:50.422743 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd05c7955_ab58_4997_847d_559aa181ea8e.slice/crio-612aed70c32da9f293cf05d7391c31ff5a1ae86af8f32bc2be09d50f17a4c189 WatchSource:0}: Error finding container 612aed70c32da9f293cf05d7391c31ff5a1ae86af8f32bc2be09d50f17a4c189: Status 404 returned error can't find the container with id 612aed70c32da9f293cf05d7391c31ff5a1ae86af8f32bc2be09d50f17a4c189 Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.515029 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.515627 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.515700 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.515781 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.515838 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:50Z","lastTransitionTime":"2025-11-24T14:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.619476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.619524 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.619534 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.619552 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.619563 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:50Z","lastTransitionTime":"2025-11-24T14:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.726167 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.726267 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.726296 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.726327 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.726355 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:50Z","lastTransitionTime":"2025-11-24T14:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.828681 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.828733 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.828743 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.828758 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.828769 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:50Z","lastTransitionTime":"2025-11-24T14:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.901739 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-x8rxd" event={"ID":"d05c7955-ab58-4997-847d-559aa181ea8e","Type":"ContainerStarted","Data":"edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.901882 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-x8rxd" event={"ID":"d05c7955-ab58-4997-847d-559aa181ea8e","Type":"ContainerStarted","Data":"612aed70c32da9f293cf05d7391c31ff5a1ae86af8f32bc2be09d50f17a4c189"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.918560 4822 generic.go:334] "Generic (PLEG): container finished" podID="45082ac5-a419-4ce2-a11b-e9a6b8178698" containerID="b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f" exitCode=0 Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.918793 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerDied","Data":"b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.921497 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.924315 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.932141 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.932198 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.932228 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.932248 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.932261 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:50Z","lastTransitionTime":"2025-11-24T14:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.939015 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.956727 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.972616 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:50 crc kubenswrapper[4822]: I1124 14:19:50.984741 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.001289 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:50Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.018999 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.035476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.035516 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.035525 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.035545 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.035555 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.038310 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.051484 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.065916 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.079279 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.091195 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.102915 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.117734 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.130793 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.140048 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.140092 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.140106 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.140125 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.140139 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.144873 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.156129 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.166247 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.176489 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.188839 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.203459 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.218606 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.232625 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.242964 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.243011 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.243024 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.243044 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.243059 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.246033 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.263297 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.273692 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.288706 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.305308 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.346588 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.346627 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.346640 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.346657 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.346671 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.363674 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.363782 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.363818 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.363859 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.363884 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364005 4822 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364059 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:59.364042529 +0000 UTC m=+36.480683026 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364475 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:19:59.364463212 +0000 UTC m=+36.481103699 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364551 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364570 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364583 4822 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364613 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:59.364602787 +0000 UTC m=+36.481243284 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364650 4822 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364676 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:59.364667319 +0000 UTC m=+36.481307806 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364724 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364735 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364745 4822 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.364769 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 14:19:59.364761722 +0000 UTC m=+36.481402219 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.449928 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.449966 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.449975 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.449989 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.449999 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.553296 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.553365 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.553388 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.553419 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.553443 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.656429 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.656486 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.656506 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.656532 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.656550 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.704258 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.704430 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.704598 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.704622 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.704730 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:19:51 crc kubenswrapper[4822]: E1124 14:19:51.704824 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.759919 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.759976 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.759998 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.760030 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.760056 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.863454 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.863520 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.863547 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.863578 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.863603 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.931701 4822 generic.go:334] "Generic (PLEG): container finished" podID="45082ac5-a419-4ce2-a11b-e9a6b8178698" containerID="7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249" exitCode=0 Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.931769 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerDied","Data":"7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.966103 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.970812 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.970859 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.970872 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.970893 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.970908 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:51Z","lastTransitionTime":"2025-11-24T14:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.981894 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:51 crc kubenswrapper[4822]: I1124 14:19:51.997812 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.010270 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.026450 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.054275 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.067098 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.072799 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.072842 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.072857 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.072876 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.072889 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:52Z","lastTransitionTime":"2025-11-24T14:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.078577 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.091020 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.105354 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.119295 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.134575 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.152504 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.167267 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.175256 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.175317 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.175334 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.175358 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.175376 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:52Z","lastTransitionTime":"2025-11-24T14:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.278153 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.278222 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.278231 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.278247 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.278256 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:52Z","lastTransitionTime":"2025-11-24T14:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.380664 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.380715 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.380729 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.380747 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.380760 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:52Z","lastTransitionTime":"2025-11-24T14:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.483384 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.483456 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.483479 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.483508 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.483527 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:52Z","lastTransitionTime":"2025-11-24T14:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.586408 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.586459 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.586486 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.586519 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.586543 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:52Z","lastTransitionTime":"2025-11-24T14:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.689839 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.689912 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.689928 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.689952 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.689971 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:52Z","lastTransitionTime":"2025-11-24T14:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.793162 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.793250 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.793264 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.793288 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.793302 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:52Z","lastTransitionTime":"2025-11-24T14:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.897116 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.897194 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.897265 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.897360 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.897442 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:52Z","lastTransitionTime":"2025-11-24T14:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.940188 4822 generic.go:334] "Generic (PLEG): container finished" podID="45082ac5-a419-4ce2-a11b-e9a6b8178698" containerID="46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87" exitCode=0 Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.940287 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerDied","Data":"46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.949381 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec"} Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.961987 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:52 crc kubenswrapper[4822]: I1124 14:19:52.980164 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:52Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.000673 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.000731 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.000749 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.000773 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.000790 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.002654 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.018336 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.031303 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.043766 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.059629 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.073224 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.089588 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.103754 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.103793 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.103804 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.103822 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.103837 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.105138 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.122629 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.134693 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.146860 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.156114 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.206560 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.206589 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.206598 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.206612 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.206622 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.310518 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.310576 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.310595 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.310618 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.310635 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.414409 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.414449 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.414459 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.414476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.414488 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.486178 4822 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.524463 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.524521 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.524537 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.524584 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.524605 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.632489 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.632552 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.632572 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.632603 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.632621 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.703542 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:53 crc kubenswrapper[4822]: E1124 14:19:53.703713 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.703542 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.704316 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:53 crc kubenswrapper[4822]: E1124 14:19:53.704343 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:19:53 crc kubenswrapper[4822]: E1124 14:19:53.704416 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.725773 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.734919 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.734976 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.734995 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.735020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.735039 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.741155 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.760697 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.776700 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.793391 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.813501 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.836308 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.838106 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.838141 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.838152 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.838167 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.838177 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.870054 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.886219 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.901231 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.915603 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.931031 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.940396 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.940446 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.940459 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.940478 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.940492 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:53Z","lastTransitionTime":"2025-11-24T14:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.947039 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.966758 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerStarted","Data":"f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219"} Nov 24 14:19:53 crc kubenswrapper[4822]: I1124 14:19:53.967380 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.005296 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.023464 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.035511 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.043947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.043988 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.043999 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.044016 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.044032 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.050233 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.071111 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.082098 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.091798 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.109174 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.122664 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.135385 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.146563 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.146705 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.146765 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.146828 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.146895 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.149727 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.160790 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.172324 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.186343 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.248922 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.248956 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.248968 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.248983 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.248994 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.352951 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.352993 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.353002 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.353017 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.353027 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.454964 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.455019 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.455033 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.455052 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.455064 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.557637 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.557700 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.557718 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.557741 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.557759 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.661098 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.661169 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.661192 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.661292 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.661318 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.764293 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.764537 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.764560 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.764589 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.764615 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.869391 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.869676 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.869696 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.869721 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.869740 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.972314 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.972377 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.972396 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.972419 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.972436 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:54Z","lastTransitionTime":"2025-11-24T14:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.977333 4822 generic.go:334] "Generic (PLEG): container finished" podID="45082ac5-a419-4ce2-a11b-e9a6b8178698" containerID="f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219" exitCode=0 Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.977414 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerDied","Data":"f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.987334 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a"} Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.987802 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:54 crc kubenswrapper[4822]: I1124 14:19:54.987857 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.001812 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.020062 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.020242 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.027886 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.045550 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.060020 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.074800 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.074853 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.074866 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.074885 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.074901 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:55Z","lastTransitionTime":"2025-11-24T14:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.076452 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.091666 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.105911 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.118352 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.129670 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.140851 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.153284 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.164426 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.177424 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.177493 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.177514 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.177546 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.177565 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:55Z","lastTransitionTime":"2025-11-24T14:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.183938 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.200881 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.213687 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.223447 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.233401 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.245525 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.258998 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.274608 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.282927 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.282991 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.283008 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.283031 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.283046 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:55Z","lastTransitionTime":"2025-11-24T14:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.294039 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.323646 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.340094 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.351301 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.362492 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.373976 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.385338 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.385542 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.385587 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.385600 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.385618 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.385629 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:55Z","lastTransitionTime":"2025-11-24T14:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.402321 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:55Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.489591 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.489667 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.489685 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.489712 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.489732 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:55Z","lastTransitionTime":"2025-11-24T14:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.592952 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.592996 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.593006 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.593022 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.593035 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:55Z","lastTransitionTime":"2025-11-24T14:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.695718 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.695805 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.695829 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.695861 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.695884 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:55Z","lastTransitionTime":"2025-11-24T14:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.704152 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.704238 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.704293 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:55 crc kubenswrapper[4822]: E1124 14:19:55.704407 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:19:55 crc kubenswrapper[4822]: E1124 14:19:55.704544 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:19:55 crc kubenswrapper[4822]: E1124 14:19:55.704673 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.798256 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.798301 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.798311 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.798326 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.798337 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:55Z","lastTransitionTime":"2025-11-24T14:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.900914 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.900948 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.900956 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.900969 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.900984 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:55Z","lastTransitionTime":"2025-11-24T14:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.995032 4822 generic.go:334] "Generic (PLEG): container finished" podID="45082ac5-a419-4ce2-a11b-e9a6b8178698" containerID="70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682" exitCode=0 Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.995163 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:19:55 crc kubenswrapper[4822]: I1124 14:19:55.996260 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerDied","Data":"70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.003020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.003332 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.003344 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.003368 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.003379 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.018353 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.036042 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.049739 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.062375 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.071815 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.083854 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.101446 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.112549 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.112591 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.112602 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.112638 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.112650 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.114595 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.166695 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.181014 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.199135 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.210586 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.221048 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.221126 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.221228 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.221271 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.221285 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.221765 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.233545 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.323848 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.324161 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.324170 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.324186 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.324197 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.428918 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.428973 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.428990 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.429014 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.429035 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.532380 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.532440 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.532463 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.532494 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.532551 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.607620 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.632549 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.639654 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.639741 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.639763 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.639793 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.639816 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.650971 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.677883 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.700591 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.717567 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.732975 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.742465 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.742548 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.742565 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.742583 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.742597 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.749818 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.768538 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.787796 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.805899 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.818158 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.828383 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.840543 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.844913 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.844959 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.844971 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.844987 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.845000 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.851926 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:56Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.947396 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.947439 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.947450 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.947470 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:56 crc kubenswrapper[4822]: I1124 14:19:56.947483 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:56Z","lastTransitionTime":"2025-11-24T14:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.001826 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.002906 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" event={"ID":"45082ac5-a419-4ce2-a11b-e9a6b8178698","Type":"ContainerStarted","Data":"e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.021585 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.032909 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.045557 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.049569 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.049605 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.049615 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.049629 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.049641 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.058778 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.067579 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.077230 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.088265 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.100698 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.114606 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.130573 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.153176 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.153249 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.153261 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.153282 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.153296 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.156937 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.170249 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.183999 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.199545 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:57Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.255587 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.255633 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.255645 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.255663 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.255676 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.358476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.358531 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.358549 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.358572 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.358591 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.461924 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.461992 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.462014 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.462042 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.462063 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.564473 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.564518 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.564529 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.564547 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.564560 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.668130 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.668245 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.668268 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.668295 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.668318 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.703941 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.703968 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.704031 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:57 crc kubenswrapper[4822]: E1124 14:19:57.704099 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:19:57 crc kubenswrapper[4822]: E1124 14:19:57.704284 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:19:57 crc kubenswrapper[4822]: E1124 14:19:57.704578 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.771736 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.771797 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.771812 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.771835 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.771848 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.874093 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.874144 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.874159 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.874176 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.874189 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.977931 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.978024 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.978049 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.978077 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:57 crc kubenswrapper[4822]: I1124 14:19:57.978096 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:57Z","lastTransitionTime":"2025-11-24T14:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.015322 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/0.log" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.020950 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a" exitCode=1 Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.021243 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.022765 4822 scope.go:117] "RemoveContainer" containerID="3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.047967 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.069451 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.081307 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.082367 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.082390 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.082413 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.082427 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:58Z","lastTransitionTime":"2025-11-24T14:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.087320 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.102631 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.125690 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:19:57.320782 6088 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:19:57.320865 6088 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 14:19:57.320882 6088 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 14:19:57.320909 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:19:57.320922 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:19:57.320943 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 14:19:57.320946 6088 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:19:57.320971 6088 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 14:19:57.320982 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:19:57.320987 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:19:57.321001 6088 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:19:57.321011 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 14:19:57.321009 6088 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 14:19:57.321083 6088 factory.go:656] Stopping watch factory\\\\nI1124 14:19:57.321133 6088 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.146450 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.171347 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.186337 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.186417 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.186442 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.186473 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.186495 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:58Z","lastTransitionTime":"2025-11-24T14:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.192580 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.211116 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.232791 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.256257 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.276676 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.290037 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.290112 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.290135 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.290169 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.290197 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:58Z","lastTransitionTime":"2025-11-24T14:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.291469 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.310972 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:58Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.394063 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.394131 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.394155 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.394187 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.394249 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:58Z","lastTransitionTime":"2025-11-24T14:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.497152 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.497197 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.497225 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.497245 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.497261 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:58Z","lastTransitionTime":"2025-11-24T14:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.599440 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.599485 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.599496 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.599516 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.599528 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:58Z","lastTransitionTime":"2025-11-24T14:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.702393 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.702434 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.702445 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.702461 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.702473 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:58Z","lastTransitionTime":"2025-11-24T14:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.804068 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.804103 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.804117 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.804132 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.804144 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:58Z","lastTransitionTime":"2025-11-24T14:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.907513 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.907556 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.907569 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.907583 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:58 crc kubenswrapper[4822]: I1124 14:19:58.907594 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:58Z","lastTransitionTime":"2025-11-24T14:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.010634 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.010724 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.010745 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.010768 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.010816 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.030366 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/0.log" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.034480 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.034802 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.063804 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.079460 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.092848 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.114385 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.115812 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.115909 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.115939 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.116013 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.116045 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.135001 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.152951 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.173875 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.193936 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:19:57.320782 6088 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:19:57.320865 6088 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 14:19:57.320882 6088 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 14:19:57.320909 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:19:57.320922 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:19:57.320943 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 14:19:57.320946 6088 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:19:57.320971 6088 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 14:19:57.320982 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:19:57.320987 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:19:57.321001 6088 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:19:57.321011 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 14:19:57.321009 6088 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 14:19:57.321083 6088 factory.go:656] Stopping watch factory\\\\nI1124 14:19:57.321133 6088 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.213472 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.218862 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.218905 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.218914 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.218930 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.218945 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.227524 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.245964 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.261073 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.275631 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.290838 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.321571 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.321614 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.321625 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.321644 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.321655 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.424877 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.424935 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.424952 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.424973 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.424987 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.461626 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.461905 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:20:15.461849898 +0000 UTC m=+52.578490375 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.462030 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.462113 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.462188 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462284 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462310 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462323 4822 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.462316 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462390 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462404 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462415 4822 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462417 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:15.462406005 +0000 UTC m=+52.579046692 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462511 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:15.462501118 +0000 UTC m=+52.579141815 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462329 4822 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462553 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:15.462545449 +0000 UTC m=+52.579186146 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462458 4822 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.462596 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:15.462589701 +0000 UTC m=+52.579230178 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.527394 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.527469 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.527493 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.527530 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.527558 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.630114 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.630160 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.630170 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.630187 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.630200 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.704383 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.704554 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.704959 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.705228 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.705371 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.705439 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.733698 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.733752 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.733785 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.733808 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.733821 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.836837 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.836881 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.836913 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.836936 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.836952 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.940339 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.940410 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.940428 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.940454 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.940472 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.958542 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.958633 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.958669 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.958701 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.958724 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.966618 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd"] Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.967785 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.971387 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.972435 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 14:19:59 crc kubenswrapper[4822]: E1124 14:19:59.979144 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.984481 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.984554 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.984573 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.984612 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.984634 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:19:59Z","lastTransitionTime":"2025-11-24T14:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:19:59 crc kubenswrapper[4822]: I1124 14:19:59.991599 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: E1124 14:20:00.000570 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:19:59Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.003367 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.004756 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.004801 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.004815 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.004835 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.004852 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.020645 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: E1124 14:20:00.026226 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.030524 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.030617 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.030632 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.030658 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.030673 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.035138 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.040103 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/1.log" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.040878 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/0.log" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.046049 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836" exitCode=1 Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.046120 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836"} Nov 24 14:20:00 crc kubenswrapper[4822]: E1124 14:20:00.046056 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.046183 4822 scope.go:117] "RemoveContainer" containerID="3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.047330 4822 scope.go:117] "RemoveContainer" containerID="d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836" Nov 24 14:20:00 crc kubenswrapper[4822]: E1124 14:20:00.047682 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.050346 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.050386 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.050399 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.050419 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.050432 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.050393 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.067039 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: E1124 14:20:00.067849 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: E1124 14:20:00.067999 4822 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.069999 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.070047 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.070063 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.070086 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.070105 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.070815 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cb351a60-3e17-4d81-815c-39bf874cf341-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.070858 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz925\" (UniqueName: \"kubernetes.io/projected/cb351a60-3e17-4d81-815c-39bf874cf341-kube-api-access-hz925\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.070928 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cb351a60-3e17-4d81-815c-39bf874cf341-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.071007 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cb351a60-3e17-4d81-815c-39bf874cf341-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.106843 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.130406 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.145787 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.158798 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.170460 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.171889 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cb351a60-3e17-4d81-815c-39bf874cf341-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.171994 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.172003 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cb351a60-3e17-4d81-815c-39bf874cf341-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.172043 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cb351a60-3e17-4d81-815c-39bf874cf341-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.172065 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz925\" (UniqueName: \"kubernetes.io/projected/cb351a60-3e17-4d81-815c-39bf874cf341-kube-api-access-hz925\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.172010 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.172666 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.172718 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.172745 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.173557 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cb351a60-3e17-4d81-815c-39bf874cf341-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.173636 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cb351a60-3e17-4d81-815c-39bf874cf341-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.177553 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cb351a60-3e17-4d81-815c-39bf874cf341-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.184940 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.186550 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz925\" (UniqueName: \"kubernetes.io/projected/cb351a60-3e17-4d81-815c-39bf874cf341-kube-api-access-hz925\") pod \"ovnkube-control-plane-749d76644c-wfbzd\" (UID: \"cb351a60-3e17-4d81-815c-39bf874cf341\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.198235 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.216674 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:19:57.320782 6088 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:19:57.320865 6088 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 14:19:57.320882 6088 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 14:19:57.320909 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:19:57.320922 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:19:57.320943 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 14:19:57.320946 6088 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:19:57.320971 6088 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 14:19:57.320982 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:19:57.320987 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:19:57.321001 6088 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:19:57.321011 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 14:19:57.321009 6088 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 14:19:57.321083 6088 factory.go:656] Stopping watch factory\\\\nI1124 14:19:57.321133 6088 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.229526 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.243039 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.254341 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.265853 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.275717 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.275774 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.275793 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.275817 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.275864 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.293163 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.298717 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:19:57.320782 6088 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:19:57.320865 6088 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 14:19:57.320882 6088 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 14:19:57.320909 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:19:57.320922 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:19:57.320943 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 14:19:57.320946 6088 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:19:57.320971 6088 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 14:19:57.320982 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:19:57.320987 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:19:57.321001 6088 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:19:57.321011 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 14:19:57.321009 6088 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 14:19:57.321083 6088 factory.go:656] Stopping watch factory\\\\nI1124 14:19:57.321133 6088 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"eduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 14:19:58.901191 6242 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-x8rxd\\\\nF1124 14:19:58.900930 6242 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.316798 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.339898 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.366082 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.382067 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.382123 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.382136 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.382154 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.382167 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.383506 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.405049 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.425859 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.439505 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.453678 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.466166 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.482419 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.489951 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.490003 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.490020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.490111 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.490131 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.497791 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:00Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.593928 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.593980 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.593992 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.594012 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.594026 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.697692 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.697754 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.697766 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.697783 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.697797 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.801369 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.801411 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.801424 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.801442 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.801453 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.904427 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.904456 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.904463 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.904478 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:00 crc kubenswrapper[4822]: I1124 14:20:00.904487 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:00Z","lastTransitionTime":"2025-11-24T14:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.006876 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.006911 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.006920 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.006935 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.006944 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.051844 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/1.log" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.057636 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" event={"ID":"cb351a60-3e17-4d81-815c-39bf874cf341","Type":"ContainerStarted","Data":"50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.057701 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" event={"ID":"cb351a60-3e17-4d81-815c-39bf874cf341","Type":"ContainerStarted","Data":"20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.057723 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" event={"ID":"cb351a60-3e17-4d81-815c-39bf874cf341","Type":"ContainerStarted","Data":"73cc0c4e543c6178d94453bee719e3f07ad5d256313572ff7878a455aa2cf4d3"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.071608 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.084533 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.100405 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.109303 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.109359 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.109379 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.109407 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.109427 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.114790 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.135242 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.155465 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.170974 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.184086 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.194993 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.209236 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.212515 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.212541 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.212548 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.212563 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.212572 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.227633 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.245103 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.268140 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.299553 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:19:57.320782 6088 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:19:57.320865 6088 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 14:19:57.320882 6088 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 14:19:57.320909 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:19:57.320922 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:19:57.320943 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 14:19:57.320946 6088 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:19:57.320971 6088 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 14:19:57.320982 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:19:57.320987 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:19:57.321001 6088 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:19:57.321011 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 14:19:57.321009 6088 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 14:19:57.321083 6088 factory.go:656] Stopping watch factory\\\\nI1124 14:19:57.321133 6088 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"eduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 14:19:58.901191 6242 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-x8rxd\\\\nF1124 14:19:58.900930 6242 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.314462 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.314541 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.314564 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.314594 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.314614 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.318902 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.417132 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.417176 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.417186 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.417224 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.417237 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.480144 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-dljmt"] Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.481001 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:01 crc kubenswrapper[4822]: E1124 14:20:01.481100 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.505509 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.520242 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.520308 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.520325 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.520368 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.520382 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.526748 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.548771 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.571975 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.586987 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.587055 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5prnq\" (UniqueName: \"kubernetes.io/projected/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-kube-api-access-5prnq\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.587696 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.607353 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.622544 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.622581 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.622593 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.622611 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.622624 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.626152 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.642290 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.660630 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.674782 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.688385 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.688428 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5prnq\" (UniqueName: \"kubernetes.io/projected/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-kube-api-access-5prnq\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:01 crc kubenswrapper[4822]: E1124 14:20:01.688629 4822 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:01 crc kubenswrapper[4822]: E1124 14:20:01.688725 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs podName:b8d0fb3c-6504-4242-b078-ee8c88e7f6f7 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:02.188703714 +0000 UTC m=+39.305344181 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs") pod "network-metrics-daemon-dljmt" (UID: "b8d0fb3c-6504-4242-b078-ee8c88e7f6f7") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.693503 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.704922 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:01 crc kubenswrapper[4822]: E1124 14:20:01.705086 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.705500 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:01 crc kubenswrapper[4822]: E1124 14:20:01.705647 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.705734 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:01 crc kubenswrapper[4822]: E1124 14:20:01.705785 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.709719 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5prnq\" (UniqueName: \"kubernetes.io/projected/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-kube-api-access-5prnq\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.719745 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:19:57.320782 6088 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:19:57.320865 6088 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 14:19:57.320882 6088 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 14:19:57.320909 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:19:57.320922 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:19:57.320943 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 14:19:57.320946 6088 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:19:57.320971 6088 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 14:19:57.320982 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:19:57.320987 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:19:57.321001 6088 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:19:57.321011 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 14:19:57.321009 6088 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 14:19:57.321083 6088 factory.go:656] Stopping watch factory\\\\nI1124 14:19:57.321133 6088 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"eduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 14:19:58.901191 6242 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-x8rxd\\\\nF1124 14:19:58.900930 6242 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.725432 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.725476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.725487 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.725504 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.725517 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.735129 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.749506 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.763795 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.774572 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:01Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.827628 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.827666 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.827680 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.827699 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.827710 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.930576 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.930655 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.930676 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.930704 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:01 crc kubenswrapper[4822]: I1124 14:20:01.930725 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:01Z","lastTransitionTime":"2025-11-24T14:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.033744 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.033804 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.033824 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.033847 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.033863 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.137387 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.137465 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.137488 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.137520 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.137542 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.194192 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:02 crc kubenswrapper[4822]: E1124 14:20:02.194427 4822 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:02 crc kubenswrapper[4822]: E1124 14:20:02.194518 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs podName:b8d0fb3c-6504-4242-b078-ee8c88e7f6f7 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:03.194493551 +0000 UTC m=+40.311134068 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs") pod "network-metrics-daemon-dljmt" (UID: "b8d0fb3c-6504-4242-b078-ee8c88e7f6f7") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.241775 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.241945 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.241969 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.241994 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.242014 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.344193 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.344284 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.344297 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.344317 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.344337 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.447575 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.447945 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.447990 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.448017 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.448041 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.551160 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.551231 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.551241 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.551259 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.551290 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.654619 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.654682 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.654701 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.654725 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.654742 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.758440 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.758497 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.758508 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.758527 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.758541 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.862385 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.862451 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.862469 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.862495 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.862520 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.966672 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.966778 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.966800 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.966878 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:02 crc kubenswrapper[4822]: I1124 14:20:02.966899 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:02Z","lastTransitionTime":"2025-11-24T14:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.069591 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.069632 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.069644 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.069659 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.069673 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:03Z","lastTransitionTime":"2025-11-24T14:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.173548 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.173611 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.173628 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.173657 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.173675 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:03Z","lastTransitionTime":"2025-11-24T14:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.207478 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:03 crc kubenswrapper[4822]: E1124 14:20:03.207674 4822 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:03 crc kubenswrapper[4822]: E1124 14:20:03.207739 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs podName:b8d0fb3c-6504-4242-b078-ee8c88e7f6f7 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:05.207722236 +0000 UTC m=+42.324362713 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs") pod "network-metrics-daemon-dljmt" (UID: "b8d0fb3c-6504-4242-b078-ee8c88e7f6f7") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.277531 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.277581 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.277598 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.277618 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.277633 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:03Z","lastTransitionTime":"2025-11-24T14:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.381576 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.381622 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.381636 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.381660 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.381674 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:03Z","lastTransitionTime":"2025-11-24T14:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.484591 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.484646 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.484664 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.484688 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.484706 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:03Z","lastTransitionTime":"2025-11-24T14:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.587713 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.587780 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.587799 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.587824 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.587841 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:03Z","lastTransitionTime":"2025-11-24T14:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.691381 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.691463 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.691481 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.691507 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.691526 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:03Z","lastTransitionTime":"2025-11-24T14:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.703685 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.703823 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.704009 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:03 crc kubenswrapper[4822]: E1124 14:20:03.704058 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:03 crc kubenswrapper[4822]: E1124 14:20:03.703854 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.703883 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:03 crc kubenswrapper[4822]: E1124 14:20:03.704498 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:03 crc kubenswrapper[4822]: E1124 14:20:03.704568 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.725945 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.739989 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.760664 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.784820 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.794467 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.794519 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.794531 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.794553 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.794569 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:03Z","lastTransitionTime":"2025-11-24T14:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.807385 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.826762 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.848333 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.865688 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.884285 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.897916 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.897992 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.898020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.898052 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.898072 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:03Z","lastTransitionTime":"2025-11-24T14:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.904537 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.924894 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.943254 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.961643 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:03 crc kubenswrapper[4822]: I1124 14:20:03.983371 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:03Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.000651 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.000704 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.000723 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.000746 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.000764 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.013894 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c939c4b62377ae47c64c4c8ea34b94e6e504323e580259ce835e7761e8b868a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:19:57.320782 6088 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:19:57.320865 6088 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 14:19:57.320882 6088 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 14:19:57.320909 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:19:57.320922 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:19:57.320943 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 14:19:57.320946 6088 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:19:57.320971 6088 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 14:19:57.320982 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:19:57.320987 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:19:57.321001 6088 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:19:57.321011 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 14:19:57.321009 6088 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 14:19:57.321083 6088 factory.go:656] Stopping watch factory\\\\nI1124 14:19:57.321133 6088 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"eduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 14:19:58.901191 6242 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-x8rxd\\\\nF1124 14:19:58.900930 6242 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:04Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.024079 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:04Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.105181 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.105285 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.105308 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.105339 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.105368 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.208538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.208588 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.208603 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.208625 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.208641 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.312260 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.312304 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.312316 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.312333 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.312346 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.415920 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.416004 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.416022 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.416050 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.416068 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.520422 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.520504 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.520528 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.520562 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.520587 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.623600 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.623687 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.623714 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.623746 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.623769 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.727103 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.727166 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.727177 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.727198 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.727228 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.830523 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.830593 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.830611 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.830638 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.830655 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.933840 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.933895 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.933909 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.933931 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:04 crc kubenswrapper[4822]: I1124 14:20:04.933948 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:04Z","lastTransitionTime":"2025-11-24T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.036841 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.036901 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.036916 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.036939 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.036955 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.140747 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.141070 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.141149 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.141242 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.141303 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.229842 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:05 crc kubenswrapper[4822]: E1124 14:20:05.230187 4822 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:05 crc kubenswrapper[4822]: E1124 14:20:05.230374 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs podName:b8d0fb3c-6504-4242-b078-ee8c88e7f6f7 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:09.230336512 +0000 UTC m=+46.346977019 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs") pod "network-metrics-daemon-dljmt" (UID: "b8d0fb3c-6504-4242-b078-ee8c88e7f6f7") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.244437 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.244520 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.244543 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.244573 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.244597 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.348748 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.348816 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.348833 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.348861 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.348881 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.452029 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.452517 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.452723 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.452947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.453133 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.557090 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.557162 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.557180 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.557241 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.557261 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.597453 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.598930 4822 scope.go:117] "RemoveContainer" containerID="d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836" Nov 24 14:20:05 crc kubenswrapper[4822]: E1124 14:20:05.599273 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.622147 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.640638 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.660688 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.661170 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.661234 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.661248 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.661271 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.661283 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.692354 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"eduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 14:19:58.901191 6242 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-x8rxd\\\\nF1124 14:19:58.900930 6242 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.703564 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:05 crc kubenswrapper[4822]: E1124 14:20:05.703927 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.703617 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:05 crc kubenswrapper[4822]: E1124 14:20:05.704768 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.703565 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:05 crc kubenswrapper[4822]: E1124 14:20:05.705073 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.703674 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:05 crc kubenswrapper[4822]: E1124 14:20:05.705388 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.711886 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.731164 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.750698 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.763947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.763978 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.764004 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.764029 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.764048 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.768954 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.787660 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.802307 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.821790 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.845751 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.860836 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.866116 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.866182 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.866198 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.866279 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.866296 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.877872 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.892560 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.906563 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:05Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.969123 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.969176 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.969193 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.969228 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:05 crc kubenswrapper[4822]: I1124 14:20:05.969248 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:05Z","lastTransitionTime":"2025-11-24T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.071959 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.072017 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.072034 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.072058 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.072077 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.174918 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.174976 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.174993 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.175015 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.175031 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.277816 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.277881 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.277901 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.277928 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.277962 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.380709 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.380745 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.380755 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.380770 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.380780 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.483507 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.483549 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.483557 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.483572 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.483584 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.585939 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.585987 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.585998 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.586017 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.586029 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.689231 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.689292 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.689312 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.689339 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.689358 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.792514 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.792816 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.792994 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.793146 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.793330 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.896258 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.896318 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.896326 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.896343 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.896354 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.999752 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.999814 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:06 crc kubenswrapper[4822]: I1124 14:20:06.999823 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:06.999847 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:06.999858 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:06Z","lastTransitionTime":"2025-11-24T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.103447 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.103920 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.104144 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.104430 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.104630 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:07Z","lastTransitionTime":"2025-11-24T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.207612 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.207687 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.207718 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.207764 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.207789 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:07Z","lastTransitionTime":"2025-11-24T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.310901 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.310968 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.310985 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.311011 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.311035 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:07Z","lastTransitionTime":"2025-11-24T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.414300 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.414353 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.414363 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.414380 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.414392 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:07Z","lastTransitionTime":"2025-11-24T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.517633 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.517712 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.517733 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.517754 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.517769 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:07Z","lastTransitionTime":"2025-11-24T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.620677 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.620997 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.621769 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.621906 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.622023 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:07Z","lastTransitionTime":"2025-11-24T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.704422 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:07 crc kubenswrapper[4822]: E1124 14:20:07.704585 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.704422 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:07 crc kubenswrapper[4822]: E1124 14:20:07.704726 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.704450 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:07 crc kubenswrapper[4822]: E1124 14:20:07.704834 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.704443 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:07 crc kubenswrapper[4822]: E1124 14:20:07.704958 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.724586 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.724635 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.724651 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.724669 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.724681 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:07Z","lastTransitionTime":"2025-11-24T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.828087 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.828167 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.828185 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.828238 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.828261 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:07Z","lastTransitionTime":"2025-11-24T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.931302 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.931377 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.931397 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.931425 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:07 crc kubenswrapper[4822]: I1124 14:20:07.931443 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:07Z","lastTransitionTime":"2025-11-24T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.034725 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.034808 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.034832 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.034868 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.034897 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.138789 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.138871 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.138895 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.138927 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.138949 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.242706 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.242776 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.242793 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.242822 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.242841 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.346500 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.346590 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.346609 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.346635 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.346655 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.449686 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.449762 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.449783 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.449809 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.449828 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.553806 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.554164 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.554348 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.554494 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.554618 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.658558 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.658609 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.658620 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.658638 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.658650 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.761703 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.761761 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.761784 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.761812 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.761834 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.865192 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.865308 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.865330 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.865360 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.865385 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.968976 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.969059 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.969077 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.969103 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:08 crc kubenswrapper[4822]: I1124 14:20:08.969123 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:08Z","lastTransitionTime":"2025-11-24T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.071707 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.071792 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.071831 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.071859 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.071882 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:09Z","lastTransitionTime":"2025-11-24T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.175451 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.175544 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.175571 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.175605 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.175629 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:09Z","lastTransitionTime":"2025-11-24T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.279624 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.279689 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.279713 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.279741 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.279763 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:09Z","lastTransitionTime":"2025-11-24T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.279891 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:09 crc kubenswrapper[4822]: E1124 14:20:09.280159 4822 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:09 crc kubenswrapper[4822]: E1124 14:20:09.280382 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs podName:b8d0fb3c-6504-4242-b078-ee8c88e7f6f7 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:17.280312851 +0000 UTC m=+54.396953368 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs") pod "network-metrics-daemon-dljmt" (UID: "b8d0fb3c-6504-4242-b078-ee8c88e7f6f7") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.383317 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.383390 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.383409 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.383432 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.383450 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:09Z","lastTransitionTime":"2025-11-24T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.487146 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.487250 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.487268 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.487297 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.487314 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:09Z","lastTransitionTime":"2025-11-24T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.590592 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.590646 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.590662 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.590685 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.590702 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:09Z","lastTransitionTime":"2025-11-24T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.694140 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.694565 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.694754 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.694910 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.695042 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:09Z","lastTransitionTime":"2025-11-24T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.703494 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.703506 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.703864 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:09 crc kubenswrapper[4822]: E1124 14:20:09.704118 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.704142 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:09 crc kubenswrapper[4822]: E1124 14:20:09.704428 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:09 crc kubenswrapper[4822]: E1124 14:20:09.704496 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:09 crc kubenswrapper[4822]: E1124 14:20:09.704624 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.798131 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.798188 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.798233 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.798260 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.798280 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:09Z","lastTransitionTime":"2025-11-24T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.901923 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.901992 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.902009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.902033 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:09 crc kubenswrapper[4822]: I1124 14:20:09.902054 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:09Z","lastTransitionTime":"2025-11-24T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.005538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.005592 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.005611 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.005632 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.005648 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.109099 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.109513 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.109646 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.109904 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.110026 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.214334 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.214418 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.214443 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.214475 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.214515 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.317818 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.317881 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.317899 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.317924 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.317943 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.328355 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.328409 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.328420 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.328439 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.328452 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: E1124 14:20:10.351332 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:10Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.356830 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.356906 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.356928 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.356958 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.356982 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: E1124 14:20:10.378448 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:10Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.384158 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.384244 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.384257 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.384280 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.384295 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: E1124 14:20:10.406422 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:10Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.411033 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.411090 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.411100 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.411120 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.411135 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: E1124 14:20:10.425410 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:10Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.430316 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.430375 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.430394 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.430417 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.430432 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: E1124 14:20:10.445922 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:10Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:10 crc kubenswrapper[4822]: E1124 14:20:10.446065 4822 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.448659 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.448716 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.448733 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.448754 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.448769 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.551985 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.552038 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.552052 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.552073 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.552091 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.654755 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.655107 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.655254 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.655360 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.655445 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.758450 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.758534 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.758558 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.758594 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.758620 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.861712 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.861773 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.861790 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.861818 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.861845 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.965436 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.965502 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.965528 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.965559 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:10 crc kubenswrapper[4822]: I1124 14:20:10.965581 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:10Z","lastTransitionTime":"2025-11-24T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.069007 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.069086 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.069110 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.069142 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.069165 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:11Z","lastTransitionTime":"2025-11-24T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.172404 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.172740 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.172971 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.173197 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.173481 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:11Z","lastTransitionTime":"2025-11-24T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.276577 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.276633 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.276650 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.276671 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.276689 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:11Z","lastTransitionTime":"2025-11-24T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.379987 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.380046 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.380066 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.380091 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.380110 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:11Z","lastTransitionTime":"2025-11-24T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.482748 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.483387 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.483419 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.483450 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.483471 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:11Z","lastTransitionTime":"2025-11-24T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.586452 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.586507 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.586516 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.586533 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.586548 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:11Z","lastTransitionTime":"2025-11-24T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.689240 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.689275 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.689284 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.689297 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.689305 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:11Z","lastTransitionTime":"2025-11-24T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.703925 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:11 crc kubenswrapper[4822]: E1124 14:20:11.704952 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.705493 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.705559 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.705679 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:11 crc kubenswrapper[4822]: E1124 14:20:11.705751 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:11 crc kubenswrapper[4822]: E1124 14:20:11.705930 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:11 crc kubenswrapper[4822]: E1124 14:20:11.710454 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.792533 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.792587 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.792601 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.792625 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.792639 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:11Z","lastTransitionTime":"2025-11-24T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.896640 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.896700 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.896712 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.896737 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:11 crc kubenswrapper[4822]: I1124 14:20:11.896750 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:11Z","lastTransitionTime":"2025-11-24T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.000538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.000606 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.000659 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.000689 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.000712 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.103555 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.103617 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.103641 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.103670 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.103693 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.207487 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.207586 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.207612 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.207645 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.207672 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.310898 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.310961 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.310981 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.311015 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.311037 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.414809 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.414920 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.414938 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.414970 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.414990 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.518123 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.518185 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.518233 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.518262 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.518284 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.621156 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.621274 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.621301 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.621333 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.621352 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.724327 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.724696 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.724824 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.725055 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.725241 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.828106 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.828156 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.828170 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.828187 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.828199 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.931626 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.931694 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.931712 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.931742 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:12 crc kubenswrapper[4822]: I1124 14:20:12.931762 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:12Z","lastTransitionTime":"2025-11-24T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.034922 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.034974 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.034990 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.035041 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.035060 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.138020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.138112 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.138135 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.138166 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.138189 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.242466 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.242523 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.242541 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.242572 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.242592 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.344971 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.345034 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.345051 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.345075 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.345093 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.448414 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.448457 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.448465 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.448480 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.448491 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.551631 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.551701 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.551726 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.551755 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.551777 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.633497 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.648956 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.654752 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.654817 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.654840 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.654868 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.654894 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.663429 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.683016 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.699173 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.703504 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.703613 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:13 crc kubenswrapper[4822]: E1124 14:20:13.703613 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.703730 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:13 crc kubenswrapper[4822]: E1124 14:20:13.703781 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.703855 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:13 crc kubenswrapper[4822]: E1124 14:20:13.705768 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:13 crc kubenswrapper[4822]: E1124 14:20:13.705863 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.720516 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.740171 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.758114 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.758152 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.758163 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.758180 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.758192 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.774685 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.826195 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.840392 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.851565 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.860504 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.860586 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.860597 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.860615 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.860625 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.865736 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.880938 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.899994 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"eduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 14:19:58.901191 6242 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-x8rxd\\\\nF1124 14:19:58.900930 6242 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.918288 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.935892 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.951025 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.963293 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.963353 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.963364 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.963385 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.963399 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:13Z","lastTransitionTime":"2025-11-24T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.964660 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.980325 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:13 crc kubenswrapper[4822]: I1124 14:20:13.999473 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:13Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.013504 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.033126 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"eduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 14:19:58.901191 6242 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-x8rxd\\\\nF1124 14:19:58.900930 6242 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.049889 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.064888 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.066394 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.066453 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.066465 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.066485 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.066498 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:14Z","lastTransitionTime":"2025-11-24T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.081141 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.093584 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.108488 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.128504 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.146483 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.167928 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.169383 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.169457 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.169486 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.169505 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.169521 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:14Z","lastTransitionTime":"2025-11-24T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.187161 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.205591 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.220952 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.237688 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.251604 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:14Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.273038 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.273293 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.273443 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.273580 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.273738 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:14Z","lastTransitionTime":"2025-11-24T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.377442 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.377735 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.377924 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.378116 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.378330 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:14Z","lastTransitionTime":"2025-11-24T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.481884 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.482297 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.482497 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.482658 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.482807 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:14Z","lastTransitionTime":"2025-11-24T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.586155 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.586281 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.586311 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.586336 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.586355 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:14Z","lastTransitionTime":"2025-11-24T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.689752 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.690089 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.690358 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.690536 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.690661 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:14Z","lastTransitionTime":"2025-11-24T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.794546 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.794982 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.795171 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.795437 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.795628 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:14Z","lastTransitionTime":"2025-11-24T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.899058 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.899103 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.899115 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.899132 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:14 crc kubenswrapper[4822]: I1124 14:20:14.899143 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:14Z","lastTransitionTime":"2025-11-24T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.002442 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.002494 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.002510 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.002534 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.002551 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.105906 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.105964 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.105987 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.106015 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.106036 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.210145 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.210196 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.210238 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.210262 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.210280 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.313846 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.314248 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.314390 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.314539 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.314681 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.417700 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.417935 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.418026 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.418111 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.418231 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.521909 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.521992 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.522016 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.522049 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.522075 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.541752 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.541879 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.541954 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:20:47.541918967 +0000 UTC m=+84.658559444 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542079 4822 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.542117 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542296 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:47.542254708 +0000 UTC m=+84.658895195 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.542486 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542298 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542568 4822 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.542575 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542595 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542624 4822 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542626 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:47.54261436 +0000 UTC m=+84.659254837 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542722 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:47.542689232 +0000 UTC m=+84.659329749 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542792 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542818 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542838 4822 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.542889 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:47.542875118 +0000 UTC m=+84.659515785 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.625124 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.625172 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.625184 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.625222 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.625236 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.703888 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.704115 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.704621 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.704718 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.704660 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.704902 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.704991 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:15 crc kubenswrapper[4822]: E1124 14:20:15.705173 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.728436 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.728498 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.728513 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.728536 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.728550 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.832067 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.832163 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.832185 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.832256 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.832282 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.936089 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.936162 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.936198 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.936279 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:15 crc kubenswrapper[4822]: I1124 14:20:15.936302 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:15Z","lastTransitionTime":"2025-11-24T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.039651 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.039709 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.039726 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.039751 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.039769 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.143448 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.143507 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.143526 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.143553 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.143573 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.246689 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.246996 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.247083 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.247170 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.247273 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.349937 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.350002 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.350020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.350044 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.350061 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.452618 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.452676 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.452688 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.452707 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.452720 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.555938 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.556006 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.556026 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.556052 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.556070 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.659120 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.659273 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.659301 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.659363 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.659389 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.762348 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.762415 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.762436 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.762468 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.762490 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.865893 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.865951 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.865970 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.865999 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.866023 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.968545 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.968610 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.968635 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.968664 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:16 crc kubenswrapper[4822]: I1124 14:20:16.968686 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:16Z","lastTransitionTime":"2025-11-24T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.071940 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.072024 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.072043 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.072069 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.072089 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.175602 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.175667 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.175685 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.175714 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.175733 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.278470 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.278512 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.278523 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.278538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.278550 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.364654 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:17 crc kubenswrapper[4822]: E1124 14:20:17.364897 4822 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:17 crc kubenswrapper[4822]: E1124 14:20:17.365005 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs podName:b8d0fb3c-6504-4242-b078-ee8c88e7f6f7 nodeName:}" failed. No retries permitted until 2025-11-24 14:20:33.364983886 +0000 UTC m=+70.481624453 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs") pod "network-metrics-daemon-dljmt" (UID: "b8d0fb3c-6504-4242-b078-ee8c88e7f6f7") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.380757 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.380788 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.380796 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.380809 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.380819 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.483696 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.483763 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.483787 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.483816 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.483837 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.587873 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.587907 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.587917 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.587932 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.587947 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.690545 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.690612 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.690636 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.690668 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.690689 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.704430 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.704496 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.704430 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:17 crc kubenswrapper[4822]: E1124 14:20:17.704635 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:17 crc kubenswrapper[4822]: E1124 14:20:17.704769 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:17 crc kubenswrapper[4822]: E1124 14:20:17.704898 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.705022 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:17 crc kubenswrapper[4822]: E1124 14:20:17.705171 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.793477 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.793546 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.793567 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.793593 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.793615 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.896464 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.896518 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.896533 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.896551 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.896718 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.999435 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.999507 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.999523 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:17 crc kubenswrapper[4822]: I1124 14:20:17.999547 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:17.999564 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:17Z","lastTransitionTime":"2025-11-24T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.103174 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.103260 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.103277 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.103301 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.103318 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:18Z","lastTransitionTime":"2025-11-24T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.210639 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.210693 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.210708 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.210730 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.210752 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:18Z","lastTransitionTime":"2025-11-24T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.315011 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.315072 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.315090 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.315140 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.315174 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:18Z","lastTransitionTime":"2025-11-24T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.419485 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.419566 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.419593 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.420144 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.420468 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:18Z","lastTransitionTime":"2025-11-24T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.525318 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.525386 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.525408 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.525435 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.525454 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:18Z","lastTransitionTime":"2025-11-24T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.628521 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.628582 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.628599 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.628625 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.628642 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:18Z","lastTransitionTime":"2025-11-24T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.705727 4822 scope.go:117] "RemoveContainer" containerID="d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.734182 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.734695 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.734982 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.735149 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.735334 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:18Z","lastTransitionTime":"2025-11-24T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.837932 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.837969 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.837980 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.837996 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.838008 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:18Z","lastTransitionTime":"2025-11-24T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.942540 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.942896 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.942911 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.942930 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:18 crc kubenswrapper[4822]: I1124 14:20:18.942942 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:18Z","lastTransitionTime":"2025-11-24T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.047060 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.047126 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.047143 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.047168 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.047186 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.133303 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/1.log" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.137194 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.137845 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.149775 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.149824 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.149839 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.149859 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.149879 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.170758 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.184757 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.209481 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.233869 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.247743 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.252850 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.252900 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.252911 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.252930 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.252942 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.262270 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.276859 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.292756 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.309696 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.323933 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.341396 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.354870 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.354903 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.354914 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.354932 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.354944 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.363262 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"eduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 14:19:58.901191 6242 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-x8rxd\\\\nF1124 14:19:58.900930 6242 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.376722 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.389009 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.404958 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.416957 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.433102 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:19Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.457925 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.457966 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.457978 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.457997 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.458009 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.561004 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.561039 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.561048 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.561061 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.561069 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.665228 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.665269 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.665284 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.665301 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.665314 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.704091 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.704152 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.704239 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:19 crc kubenswrapper[4822]: E1124 14:20:19.704254 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.704284 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:19 crc kubenswrapper[4822]: E1124 14:20:19.704387 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:19 crc kubenswrapper[4822]: E1124 14:20:19.704447 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:19 crc kubenswrapper[4822]: E1124 14:20:19.704589 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.768053 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.768096 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.768106 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.768121 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.768135 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.871615 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.871669 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.871689 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.871711 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.871728 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.974679 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.974744 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.974764 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.974789 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:19 crc kubenswrapper[4822]: I1124 14:20:19.974807 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:19Z","lastTransitionTime":"2025-11-24T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.078271 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.078327 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.078339 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.078358 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.078371 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.144326 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/2.log" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.145462 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/1.log" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.150032 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9" exitCode=1 Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.150095 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9"} Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.150156 4822 scope.go:117] "RemoveContainer" containerID="d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.151475 4822 scope.go:117] "RemoveContainer" containerID="9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9" Nov 24 14:20:20 crc kubenswrapper[4822]: E1124 14:20:20.151821 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.180898 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.181023 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.181053 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.181088 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.181119 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.190188 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d878bd6b94d93bf78b3d57861b6e25ada3e24ed91f5350b69d165a005b5fd836\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"message\\\":\\\"eduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 14:19:58.901191 6242 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-x8rxd\\\\nF1124 14:19:58.900930 6242 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:19Z\\\",\\\"message\\\":\\\"(0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:20:19.685746 6517 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:20:19.685774 6517 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:20:19.685815 6517 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:20:19.685850 6517 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:20:19.686345 6517 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:20:19.686442 6517 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:20:19.686483 6517 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:20:19.686637 6517 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 14:20:19.686649 6517 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 14:20:19.686672 6517 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 14:20:19.686694 6517 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 14:20:19.686719 6517 factory.go:656] Stopping watch factory\\\\nI1124 14:20:19.686740 6517 ovnkube.go:599] Stopped ovnkube\\\\nI1124 14:20:19.686758 6517 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 14:20:19.686812 6517 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 14:20:19.686970 6517 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.205721 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.223338 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.238094 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.256908 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.273790 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.284425 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.284502 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.284529 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.284563 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.284588 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.298705 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.315129 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.335778 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.352929 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.369367 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.387593 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.387640 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.387650 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.387668 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.387681 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.390696 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.409523 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.428565 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.449343 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.462333 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.475071 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.490087 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.490135 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.490145 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.490161 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.490175 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.593616 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.593677 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.593689 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.593711 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.593725 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.673050 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.673134 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.673159 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.673192 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.673247 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: E1124 14:20:20.699903 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.704500 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.704550 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.704569 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.704593 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.704612 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: E1124 14:20:20.720888 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.725350 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.725391 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.725402 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.725419 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.725430 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: E1124 14:20:20.743684 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.748179 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.748510 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.748533 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.748550 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.748562 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: E1124 14:20:20.769063 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.775199 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.775265 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.775285 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.775310 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.775328 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: E1124 14:20:20.796181 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:20Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:20 crc kubenswrapper[4822]: E1124 14:20:20.797027 4822 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.799423 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.799667 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.800018 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.800350 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.800650 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.904272 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.904318 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.904335 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.904361 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:20 crc kubenswrapper[4822]: I1124 14:20:20.904379 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:20Z","lastTransitionTime":"2025-11-24T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.006941 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.007404 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.007620 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.007754 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.007883 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.111641 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.112061 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.112260 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.112445 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.112602 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.155704 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/2.log" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.160775 4822 scope.go:117] "RemoveContainer" containerID="9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9" Nov 24 14:20:21 crc kubenswrapper[4822]: E1124 14:20:21.161057 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.177771 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.194102 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.215344 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.216305 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.216371 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.216390 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.216415 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.216432 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.236622 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.257749 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.275453 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.301814 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.319068 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.319115 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.319126 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.319140 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.319152 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.337126 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:19Z\\\",\\\"message\\\":\\\"(0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:20:19.685746 6517 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:20:19.685774 6517 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:20:19.685815 6517 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:20:19.685850 6517 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:20:19.686345 6517 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:20:19.686442 6517 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:20:19.686483 6517 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:20:19.686637 6517 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 14:20:19.686649 6517 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 14:20:19.686672 6517 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 14:20:19.686694 6517 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 14:20:19.686719 6517 factory.go:656] Stopping watch factory\\\\nI1124 14:20:19.686740 6517 ovnkube.go:599] Stopped ovnkube\\\\nI1124 14:20:19.686758 6517 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 14:20:19.686812 6517 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 14:20:19.686970 6517 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.355547 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.374090 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.390891 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.405733 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.422142 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.422242 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.422269 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.422300 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.422321 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.425966 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.445075 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.464132 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.483485 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.499897 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:21Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.525575 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.525601 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.525610 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.525626 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.525635 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.627834 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.628130 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.628250 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.628342 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.628427 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.703970 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.703971 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.704072 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:21 crc kubenswrapper[4822]: E1124 14:20:21.704314 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.704001 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:21 crc kubenswrapper[4822]: E1124 14:20:21.704468 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:21 crc kubenswrapper[4822]: E1124 14:20:21.704604 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:21 crc kubenswrapper[4822]: E1124 14:20:21.704705 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.735394 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.735461 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.735479 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.735502 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.735521 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.838576 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.838624 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.838643 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.838664 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.838681 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.942621 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.942679 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.942695 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.942717 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:21 crc kubenswrapper[4822]: I1124 14:20:21.942733 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:21Z","lastTransitionTime":"2025-11-24T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.044747 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.045068 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.045152 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.045292 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.045405 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.148599 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.148692 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.148720 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.148748 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.148771 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.252055 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.252115 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.252133 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.252160 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.252181 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.355483 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.355529 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.355539 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.355561 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.355576 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.458022 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.458404 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.458564 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.458676 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.458785 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.562188 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.562251 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.562259 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.562279 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.562289 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.666017 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.666086 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.666110 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.666137 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.666161 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.768972 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.769319 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.769435 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.769527 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.769618 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.872389 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.872962 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.873056 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.873141 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.873251 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.976984 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.977371 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.977524 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.977682 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:22 crc kubenswrapper[4822]: I1124 14:20:22.977820 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:22Z","lastTransitionTime":"2025-11-24T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.080639 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.080701 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.080718 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.080743 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.080762 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:23Z","lastTransitionTime":"2025-11-24T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.183565 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.183657 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.183675 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.183707 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.183728 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:23Z","lastTransitionTime":"2025-11-24T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.286361 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.286724 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.286873 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.287021 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.287168 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:23Z","lastTransitionTime":"2025-11-24T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.390099 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.390155 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.390172 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.390194 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.390260 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:23Z","lastTransitionTime":"2025-11-24T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.493402 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.493771 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.493995 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.494236 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.494455 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:23Z","lastTransitionTime":"2025-11-24T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.597401 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.597529 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.597560 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.597590 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.597614 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:23Z","lastTransitionTime":"2025-11-24T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.701007 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.701078 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.701098 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.701122 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.701141 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:23Z","lastTransitionTime":"2025-11-24T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.703476 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.703557 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.703566 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.703498 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:23 crc kubenswrapper[4822]: E1124 14:20:23.703697 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:23 crc kubenswrapper[4822]: E1124 14:20:23.703759 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:23 crc kubenswrapper[4822]: E1124 14:20:23.703830 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:23 crc kubenswrapper[4822]: E1124 14:20:23.704024 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.726641 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.750341 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.771243 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.786810 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.804029 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.804070 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.804079 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.804094 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.804105 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:23Z","lastTransitionTime":"2025-11-24T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.804155 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.825616 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.841572 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.859803 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.876650 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.894296 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.907331 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.907375 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.907388 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.907403 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.907414 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:23Z","lastTransitionTime":"2025-11-24T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.915881 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:19Z\\\",\\\"message\\\":\\\"(0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:20:19.685746 6517 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:20:19.685774 6517 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:20:19.685815 6517 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:20:19.685850 6517 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:20:19.686345 6517 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:20:19.686442 6517 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:20:19.686483 6517 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:20:19.686637 6517 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 14:20:19.686649 6517 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 14:20:19.686672 6517 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 14:20:19.686694 6517 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 14:20:19.686719 6517 factory.go:656] Stopping watch factory\\\\nI1124 14:20:19.686740 6517 ovnkube.go:599] Stopped ovnkube\\\\nI1124 14:20:19.686758 6517 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 14:20:19.686812 6517 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 14:20:19.686970 6517 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.929337 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.941955 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.953983 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.967949 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:23 crc kubenswrapper[4822]: I1124 14:20:23.985826 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.000903 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:23Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.010350 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.010409 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.010419 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.010437 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.010449 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.118844 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.118957 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.118980 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.119009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.119032 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.221972 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.222034 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.222047 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.222066 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.222080 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.325438 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.325520 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.325540 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.325560 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.325575 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.428300 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.428356 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.428371 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.428395 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.428413 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.530854 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.530894 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.530903 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.530917 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.530928 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.634123 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.634184 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.634245 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.634280 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.634303 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.737294 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.737373 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.737393 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.737415 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.737431 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.841618 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.841668 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.841686 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.841709 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.841726 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.944390 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.944425 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.944440 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.944460 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:24 crc kubenswrapper[4822]: I1124 14:20:24.944474 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:24Z","lastTransitionTime":"2025-11-24T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.047036 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.047095 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.047109 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.047128 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.047142 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.149918 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.149997 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.150018 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.150043 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.150061 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.252495 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.252556 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.252575 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.252599 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.252617 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.355309 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.355379 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.355399 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.355423 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.355440 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.458091 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.458175 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.458201 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.458270 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.458292 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.561543 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.561628 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.561652 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.561681 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.561700 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.664240 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.664299 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.664314 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.664335 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.664348 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.704028 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.704061 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.704107 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.704128 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:25 crc kubenswrapper[4822]: E1124 14:20:25.704258 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:25 crc kubenswrapper[4822]: E1124 14:20:25.704403 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:25 crc kubenswrapper[4822]: E1124 14:20:25.704505 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:25 crc kubenswrapper[4822]: E1124 14:20:25.704572 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.767748 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.767915 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.767940 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.767961 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.767978 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.871097 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.871165 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.871182 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.871239 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.871261 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.974691 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.974730 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.974738 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.974750 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:25 crc kubenswrapper[4822]: I1124 14:20:25.974760 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:25Z","lastTransitionTime":"2025-11-24T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.077821 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.077912 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.077933 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.077970 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.077995 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:26Z","lastTransitionTime":"2025-11-24T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.180266 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.180353 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.180363 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.180376 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.180389 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:26Z","lastTransitionTime":"2025-11-24T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.283062 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.283134 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.283155 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.283180 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.283200 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:26Z","lastTransitionTime":"2025-11-24T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.386375 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.386440 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.386459 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.386487 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.386512 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:26Z","lastTransitionTime":"2025-11-24T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.489790 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.489832 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.489843 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.489859 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.489872 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:26Z","lastTransitionTime":"2025-11-24T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.592748 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.592821 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.592835 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.592888 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.592904 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:26Z","lastTransitionTime":"2025-11-24T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.695568 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.695622 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.695634 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.695654 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.695678 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:26Z","lastTransitionTime":"2025-11-24T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.799390 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.799466 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.799488 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.799521 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.799544 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:26Z","lastTransitionTime":"2025-11-24T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.903471 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.903538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.903559 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.903587 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:26 crc kubenswrapper[4822]: I1124 14:20:26.903607 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:26Z","lastTransitionTime":"2025-11-24T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.006512 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.006590 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.006611 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.006641 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.006665 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.110054 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.110116 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.110129 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.110147 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.110160 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.213710 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.213775 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.213791 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.213816 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.213833 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.316691 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.316786 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.316809 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.317417 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.317476 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.421718 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.421747 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.421755 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.421770 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.421803 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.524949 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.524998 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.525010 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.525029 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.525043 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.627824 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.627892 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.627912 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.627937 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.627956 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.704250 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.704309 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.704337 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:27 crc kubenswrapper[4822]: E1124 14:20:27.704445 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.704727 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:27 crc kubenswrapper[4822]: E1124 14:20:27.704750 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:27 crc kubenswrapper[4822]: E1124 14:20:27.704801 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:27 crc kubenswrapper[4822]: E1124 14:20:27.704851 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.731731 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.731792 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.731810 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.731837 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.731857 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.835132 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.835184 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.835238 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.835264 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.835282 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.938417 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.938468 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.938481 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.938498 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:27 crc kubenswrapper[4822]: I1124 14:20:27.938510 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:27Z","lastTransitionTime":"2025-11-24T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.041731 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.041789 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.041804 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.041823 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.041837 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.144571 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.144627 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.144639 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.144657 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.144674 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.247243 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.247304 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.247324 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.247346 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.247362 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.351199 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.351260 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.351269 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.351284 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.351294 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.453705 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.453752 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.453762 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.453776 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.453787 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.556388 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.556451 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.556473 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.556508 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.556526 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.658892 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.658933 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.658947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.658963 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.658982 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.761346 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.761428 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.761452 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.761483 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.761505 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.864127 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.864454 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.864538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.864644 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.864730 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.968085 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.968128 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.968139 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.968155 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:28 crc kubenswrapper[4822]: I1124 14:20:28.968166 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:28Z","lastTransitionTime":"2025-11-24T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.070123 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.070170 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.070181 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.070223 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.070237 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.172461 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.172538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.172559 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.172583 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.172601 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.275177 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.275260 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.275278 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.275304 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.275322 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.377264 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.377310 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.377321 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.377336 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.377345 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.479911 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.479957 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.479966 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.479980 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.479992 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.582287 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.582357 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.582372 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.582395 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.582412 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.685912 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.686021 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.686075 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.686101 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.686129 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.704001 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.704058 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.704011 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.704001 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:29 crc kubenswrapper[4822]: E1124 14:20:29.704173 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:29 crc kubenswrapper[4822]: E1124 14:20:29.704269 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:29 crc kubenswrapper[4822]: E1124 14:20:29.704331 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:29 crc kubenswrapper[4822]: E1124 14:20:29.704490 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.788960 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.789040 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.789058 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.789085 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.789105 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.894110 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.894183 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.894194 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.894232 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.894247 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.997013 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.997061 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.997072 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.997091 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:29 crc kubenswrapper[4822]: I1124 14:20:29.997103 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:29Z","lastTransitionTime":"2025-11-24T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.099800 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.099849 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.099860 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.099878 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.099893 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.202643 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.202713 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.202730 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.202754 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.202770 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.305599 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.305642 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.305654 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.305671 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.305681 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.408794 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.409238 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.409445 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.409641 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.409812 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.511867 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.511894 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.511905 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.511919 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.511929 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.614407 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.614460 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.614474 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.614494 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.614508 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.717454 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.717816 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.717950 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.718074 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.718255 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.810996 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.811055 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.811073 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.811094 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.811109 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: E1124 14:20:30.832355 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:30Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.837254 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.837320 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.837333 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.837351 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.837361 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: E1124 14:20:30.853137 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:30Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.857945 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.858192 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.858411 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.858603 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.858792 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: E1124 14:20:30.873643 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:30Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.878231 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.878311 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.878332 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.878358 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.878371 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: E1124 14:20:30.892136 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:30Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.896571 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.896736 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.896843 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.896922 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.896994 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:30 crc kubenswrapper[4822]: E1124 14:20:30.909093 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:30Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:30 crc kubenswrapper[4822]: E1124 14:20:30.909489 4822 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.911588 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.911642 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.911656 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.911678 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:30 crc kubenswrapper[4822]: I1124 14:20:30.911694 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:30Z","lastTransitionTime":"2025-11-24T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.014755 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.014821 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.014839 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.014865 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.014887 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.117564 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.117603 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.117617 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.117633 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.117644 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.220563 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.220607 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.220617 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.220634 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.220646 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.323250 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.323596 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.323690 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.323782 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.323862 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.427986 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.428386 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.428524 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.428664 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.428797 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.531393 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.531656 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.531723 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.531791 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.531860 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.634689 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.634984 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.635085 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.635186 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.635292 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.704447 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.704468 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.704498 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.704548 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:31 crc kubenswrapper[4822]: E1124 14:20:31.704996 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.705225 4822 scope.go:117] "RemoveContainer" containerID="9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9" Nov 24 14:20:31 crc kubenswrapper[4822]: E1124 14:20:31.705381 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:20:31 crc kubenswrapper[4822]: E1124 14:20:31.705460 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:31 crc kubenswrapper[4822]: E1124 14:20:31.705699 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:31 crc kubenswrapper[4822]: E1124 14:20:31.705771 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.737532 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.737604 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.737624 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.737646 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.737662 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.839947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.839987 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.839995 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.840022 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.840033 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.942327 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.942369 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.942378 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.942395 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:31 crc kubenswrapper[4822]: I1124 14:20:31.942406 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:31Z","lastTransitionTime":"2025-11-24T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.044623 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.044655 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.044663 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.044678 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.044688 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.147701 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.147760 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.147777 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.147805 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.147829 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.251356 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.251421 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.251441 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.251468 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.251489 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.354414 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.354458 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.354470 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.354487 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.354497 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.457485 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.457525 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.457538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.457553 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.457564 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.559713 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.559767 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.559780 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.559798 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.559811 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.662111 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.662160 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.662170 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.662186 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.662215 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.764947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.764989 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.765001 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.765021 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.765035 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.866862 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.866894 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.866903 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.866918 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.866928 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.969345 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.969388 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.969398 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.969412 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:32 crc kubenswrapper[4822]: I1124 14:20:32.969425 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:32Z","lastTransitionTime":"2025-11-24T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.072396 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.072439 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.072449 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.072465 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.072477 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.175082 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.175119 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.175127 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.175141 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.175150 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.277367 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.277425 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.277434 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.277449 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.277460 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.380567 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.380613 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.380622 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.380636 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.380650 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.439743 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:33 crc kubenswrapper[4822]: E1124 14:20:33.440027 4822 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:33 crc kubenswrapper[4822]: E1124 14:20:33.440124 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs podName:b8d0fb3c-6504-4242-b078-ee8c88e7f6f7 nodeName:}" failed. No retries permitted until 2025-11-24 14:21:05.440093677 +0000 UTC m=+102.556734194 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs") pod "network-metrics-daemon-dljmt" (UID: "b8d0fb3c-6504-4242-b078-ee8c88e7f6f7") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.483895 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.483966 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.483978 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.484020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.484035 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.587269 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.587331 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.587347 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.587370 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.587388 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.689971 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.690024 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.690039 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.690058 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.690071 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.703935 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.703989 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.703989 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.704148 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:33 crc kubenswrapper[4822]: E1124 14:20:33.704149 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:33 crc kubenswrapper[4822]: E1124 14:20:33.704332 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:33 crc kubenswrapper[4822]: E1124 14:20:33.704479 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:33 crc kubenswrapper[4822]: E1124 14:20:33.704666 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.719914 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.730879 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.742922 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.755996 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.774047 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.792298 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.792332 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.792342 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.792356 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.792367 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.797285 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:19Z\\\",\\\"message\\\":\\\"(0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:20:19.685746 6517 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:20:19.685774 6517 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:20:19.685815 6517 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:20:19.685850 6517 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:20:19.686345 6517 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:20:19.686442 6517 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:20:19.686483 6517 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:20:19.686637 6517 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 14:20:19.686649 6517 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 14:20:19.686672 6517 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 14:20:19.686694 6517 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 14:20:19.686719 6517 factory.go:656] Stopping watch factory\\\\nI1124 14:20:19.686740 6517 ovnkube.go:599] Stopped ovnkube\\\\nI1124 14:20:19.686758 6517 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 14:20:19.686812 6517 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 14:20:19.686970 6517 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.812785 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.826618 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.838975 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.850058 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.861475 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.875165 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.889253 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.894680 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.894731 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.894744 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.894764 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.894778 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.903883 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.925240 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.943455 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.956317 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:33Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.997758 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.997795 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.997803 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.997819 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:33 crc kubenswrapper[4822]: I1124 14:20:33.997829 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:33Z","lastTransitionTime":"2025-11-24T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.100669 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.100707 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.100717 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.100738 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.100748 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:34Z","lastTransitionTime":"2025-11-24T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.202847 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.202915 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.202929 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.202946 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.202958 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:34Z","lastTransitionTime":"2025-11-24T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.305674 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.305707 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.305716 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.305729 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.305739 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:34Z","lastTransitionTime":"2025-11-24T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.408699 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.408742 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.408753 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.408772 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.408783 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:34Z","lastTransitionTime":"2025-11-24T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.511472 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.511513 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.511521 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.511536 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.511549 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:34Z","lastTransitionTime":"2025-11-24T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.614715 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.614779 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.614801 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.614831 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.614853 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:34Z","lastTransitionTime":"2025-11-24T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.717656 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.717720 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.717741 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.717768 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.717791 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:34Z","lastTransitionTime":"2025-11-24T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.819990 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.820030 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.820040 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.820055 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.820065 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:34Z","lastTransitionTime":"2025-11-24T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.921572 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.921628 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.921641 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.921659 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:34 crc kubenswrapper[4822]: I1124 14:20:34.921676 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:34Z","lastTransitionTime":"2025-11-24T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.024385 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.024438 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.024450 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.024473 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.024486 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.127576 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.127635 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.127653 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.127676 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.127697 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.205355 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/0.log" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.205413 4822 generic.go:334] "Generic (PLEG): container finished" podID="92c230aa-aed2-4b07-82bd-98b1d2ffa456" containerID="385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16" exitCode=1 Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.205445 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-d6rbt" event={"ID":"92c230aa-aed2-4b07-82bd-98b1d2ffa456","Type":"ContainerDied","Data":"385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.205848 4822 scope.go:117] "RemoveContainer" containerID="385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.230178 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.230445 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.230458 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.230478 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.230494 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.235781 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.250181 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.264748 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:34Z\\\",\\\"message\\\":\\\"2025-11-24T14:19:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6\\\\n2025-11-24T14:19:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6 to /host/opt/cni/bin/\\\\n2025-11-24T14:19:49Z [verbose] multus-daemon started\\\\n2025-11-24T14:19:49Z [verbose] Readiness Indicator file check\\\\n2025-11-24T14:20:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.285454 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.300547 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.315260 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.329579 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.333272 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.333298 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.333319 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.333333 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.333343 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.342087 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.354423 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.368940 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.379537 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.394720 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.411677 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:19Z\\\",\\\"message\\\":\\\"(0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:20:19.685746 6517 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:20:19.685774 6517 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:20:19.685815 6517 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:20:19.685850 6517 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:20:19.686345 6517 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:20:19.686442 6517 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:20:19.686483 6517 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:20:19.686637 6517 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 14:20:19.686649 6517 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 14:20:19.686672 6517 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 14:20:19.686694 6517 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 14:20:19.686719 6517 factory.go:656] Stopping watch factory\\\\nI1124 14:20:19.686740 6517 ovnkube.go:599] Stopped ovnkube\\\\nI1124 14:20:19.686758 6517 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 14:20:19.686812 6517 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 14:20:19.686970 6517 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.422948 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.433463 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.435196 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.435242 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.435252 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.435267 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.435277 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.445815 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.457084 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:35Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.538051 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.538097 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.538115 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.538138 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.538156 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.640800 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.640842 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.640853 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.640882 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.640892 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.704006 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.704004 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:35 crc kubenswrapper[4822]: E1124 14:20:35.704827 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.704086 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:35 crc kubenswrapper[4822]: E1124 14:20:35.704929 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.704091 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:35 crc kubenswrapper[4822]: E1124 14:20:35.705069 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:35 crc kubenswrapper[4822]: E1124 14:20:35.705466 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.743865 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.743944 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.743963 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.743987 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.744006 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.846965 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.847381 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.847502 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.847647 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.848014 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.951165 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.951254 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.951269 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.951291 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:35 crc kubenswrapper[4822]: I1124 14:20:35.951304 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:35Z","lastTransitionTime":"2025-11-24T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.053012 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.053048 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.053056 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.053069 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.053079 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.155606 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.155656 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.155673 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.155697 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.155713 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.210501 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/0.log" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.210596 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-d6rbt" event={"ID":"92c230aa-aed2-4b07-82bd-98b1d2ffa456","Type":"ContainerStarted","Data":"6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.224539 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.240761 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.253850 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.257945 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.257968 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.257978 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.257991 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.258000 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.269802 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.282231 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:34Z\\\",\\\"message\\\":\\\"2025-11-24T14:19:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6\\\\n2025-11-24T14:19:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6 to /host/opt/cni/bin/\\\\n2025-11-24T14:19:49Z [verbose] multus-daemon started\\\\n2025-11-24T14:19:49Z [verbose] Readiness Indicator file check\\\\n2025-11-24T14:20:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.296275 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.313025 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.327317 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.341186 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.358978 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.361007 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.361086 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.361114 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.361146 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.361173 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.380644 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.404784 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.422200 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.444767 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.464466 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.464513 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.464529 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.464552 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.464570 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.469028 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:19Z\\\",\\\"message\\\":\\\"(0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:20:19.685746 6517 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:20:19.685774 6517 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:20:19.685815 6517 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:20:19.685850 6517 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:20:19.686345 6517 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:20:19.686442 6517 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:20:19.686483 6517 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:20:19.686637 6517 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 14:20:19.686649 6517 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 14:20:19.686672 6517 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 14:20:19.686694 6517 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 14:20:19.686719 6517 factory.go:656] Stopping watch factory\\\\nI1124 14:20:19.686740 6517 ovnkube.go:599] Stopped ovnkube\\\\nI1124 14:20:19.686758 6517 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 14:20:19.686812 6517 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 14:20:19.686970 6517 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.485575 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.501731 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:36Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.567468 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.567511 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.567520 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.567535 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.567546 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.670195 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.670455 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.670524 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.670617 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.670694 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.773492 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.773696 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.773772 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.773840 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.773904 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.876169 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.876257 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.876278 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.876300 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.876315 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.978348 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.978542 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.978628 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.978812 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:36 crc kubenswrapper[4822]: I1124 14:20:36.978927 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:36Z","lastTransitionTime":"2025-11-24T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.082123 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.082173 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.082186 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.082218 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.082417 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:37Z","lastTransitionTime":"2025-11-24T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.185145 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.185500 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.185652 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.185794 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.185910 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:37Z","lastTransitionTime":"2025-11-24T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.288500 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.289131 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.289265 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.289379 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.289482 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:37Z","lastTransitionTime":"2025-11-24T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.392229 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.392273 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.392282 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.392302 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.392312 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:37Z","lastTransitionTime":"2025-11-24T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.495558 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.495815 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.495940 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.496016 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.496079 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:37Z","lastTransitionTime":"2025-11-24T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.598009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.598299 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.598372 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.598461 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.598541 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:37Z","lastTransitionTime":"2025-11-24T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.701770 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.702115 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.702266 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.702394 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.702505 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:37Z","lastTransitionTime":"2025-11-24T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.704089 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.704130 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.704184 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.704183 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:37 crc kubenswrapper[4822]: E1124 14:20:37.704437 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:37 crc kubenswrapper[4822]: E1124 14:20:37.704609 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:37 crc kubenswrapper[4822]: E1124 14:20:37.704640 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:37 crc kubenswrapper[4822]: E1124 14:20:37.704735 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.718708 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.805670 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.805741 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.805764 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.805795 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.805818 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:37Z","lastTransitionTime":"2025-11-24T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.908792 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.908858 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.908875 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.908899 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:37 crc kubenswrapper[4822]: I1124 14:20:37.908917 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:37Z","lastTransitionTime":"2025-11-24T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.011858 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.011895 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.011907 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.011925 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.011937 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.114734 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.114794 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.114803 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.114822 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.114832 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.217624 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.217669 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.217679 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.217692 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.217702 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.320825 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.320875 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.320887 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.320910 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.320924 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.423762 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.423820 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.423837 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.423864 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.423883 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.526887 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.526960 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.526985 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.527018 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.527046 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.630635 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.630700 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.630719 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.630745 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.630765 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.741007 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.741088 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.741100 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.741118 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.741131 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.844181 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.844299 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.844321 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.844352 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.844376 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.947629 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.947685 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.947701 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.947725 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:38 crc kubenswrapper[4822]: I1124 14:20:38.947745 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:38Z","lastTransitionTime":"2025-11-24T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.050885 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.050946 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.050964 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.050989 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.051013 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.154719 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.154839 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.154897 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.154928 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.154947 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.258091 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.258180 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.258198 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.258287 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.258307 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.362183 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.362243 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.362255 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.362273 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.362287 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.465419 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.465469 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.465482 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.465500 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.465511 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.568942 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.569009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.569027 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.569052 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.569074 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.673758 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.673852 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.673877 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.673908 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.673937 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.704470 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.704470 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:39 crc kubenswrapper[4822]: E1124 14:20:39.704685 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.704809 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.704477 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:39 crc kubenswrapper[4822]: E1124 14:20:39.704883 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:39 crc kubenswrapper[4822]: E1124 14:20:39.705014 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:39 crc kubenswrapper[4822]: E1124 14:20:39.705130 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.777560 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.777639 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.777664 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.777696 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.777724 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.880427 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.880483 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.880499 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.880523 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.880541 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.983702 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.983812 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.983831 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.983861 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:39 crc kubenswrapper[4822]: I1124 14:20:39.983879 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:39Z","lastTransitionTime":"2025-11-24T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.086008 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.086078 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.086095 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.086120 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.086141 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:40Z","lastTransitionTime":"2025-11-24T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.189915 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.189985 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.190007 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.190038 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.190058 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:40Z","lastTransitionTime":"2025-11-24T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.293281 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.293351 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.293369 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.293393 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.293412 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:40Z","lastTransitionTime":"2025-11-24T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.397003 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.397064 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.397081 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.397117 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.397137 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:40Z","lastTransitionTime":"2025-11-24T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.500847 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.500896 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.500909 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.500927 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.500943 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:40Z","lastTransitionTime":"2025-11-24T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.604286 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.604355 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.604378 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.604590 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.604616 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:40Z","lastTransitionTime":"2025-11-24T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.708990 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.709048 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.709065 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.709084 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.709103 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:40Z","lastTransitionTime":"2025-11-24T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.812404 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.812491 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.812522 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.812554 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.812578 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:40Z","lastTransitionTime":"2025-11-24T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.915688 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.915760 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.915781 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.915805 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:40 crc kubenswrapper[4822]: I1124 14:20:40.915824 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:40Z","lastTransitionTime":"2025-11-24T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.003001 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.003069 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.003086 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.003130 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.003166 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.025585 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:41Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.031982 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.032060 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.032084 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.032115 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.032137 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.050561 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:41Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.062005 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.062065 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.062084 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.062111 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.062131 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.082940 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:41Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.088644 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.088708 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.088733 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.088767 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.088790 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.106698 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:41Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.112084 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.112140 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.112157 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.112185 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.112225 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.142428 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:41Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.142690 4822 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.144800 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.144867 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.144891 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.144923 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.144948 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.248021 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.248086 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.248103 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.248129 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.248147 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.351875 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.351947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.351967 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.351996 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.352017 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.455489 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.455549 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.455567 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.455596 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.455614 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.559011 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.559070 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.559087 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.559112 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.559131 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.662116 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.662192 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.662270 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.662303 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.662327 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.704182 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.704404 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.704558 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.704666 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.704629 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.705123 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.705355 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:41 crc kubenswrapper[4822]: E1124 14:20:41.705454 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.720659 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.765456 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.765510 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.765525 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.765543 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.765557 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.868621 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.868688 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.868710 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.868739 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.868758 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.971823 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.971882 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.971900 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.971923 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:41 crc kubenswrapper[4822]: I1124 14:20:41.971944 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:41Z","lastTransitionTime":"2025-11-24T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.074968 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.075047 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.075067 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.075093 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.075112 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:42Z","lastTransitionTime":"2025-11-24T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.178592 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.178681 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.178712 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.178748 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.178770 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:42Z","lastTransitionTime":"2025-11-24T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.281662 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.281757 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.281777 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.281805 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.281823 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:42Z","lastTransitionTime":"2025-11-24T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.384703 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.385160 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.385283 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.385389 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.385494 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:42Z","lastTransitionTime":"2025-11-24T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.488574 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.489681 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.489845 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.489996 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.490116 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:42Z","lastTransitionTime":"2025-11-24T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.593890 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.593956 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.593974 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.594004 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.594022 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:42Z","lastTransitionTime":"2025-11-24T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.697323 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.697388 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.697411 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.697440 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.697462 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:42Z","lastTransitionTime":"2025-11-24T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.800340 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.800450 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.800469 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.800491 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.800509 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:42Z","lastTransitionTime":"2025-11-24T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.904113 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.904161 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.904177 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.904199 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:42 crc kubenswrapper[4822]: I1124 14:20:42.904250 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:42Z","lastTransitionTime":"2025-11-24T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.007606 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.007673 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.007690 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.007716 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.007734 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.111389 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.111466 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.111488 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.111517 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.111536 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.215033 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.215094 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.215111 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.215130 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.215143 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.318062 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.318125 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.318144 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.318170 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.318190 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.420826 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.421291 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.421531 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.421776 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.422011 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.524931 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.524969 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.524979 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.524993 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.525003 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.628431 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.629007 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.629036 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.629068 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.629092 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.703877 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.703925 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.703942 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:43 crc kubenswrapper[4822]: E1124 14:20:43.704017 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.704066 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:43 crc kubenswrapper[4822]: E1124 14:20:43.704287 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:43 crc kubenswrapper[4822]: E1124 14:20:43.704318 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:43 crc kubenswrapper[4822]: E1124 14:20:43.704371 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.726938 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.731302 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.731341 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.731354 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.731376 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.731389 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.753494 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:19Z\\\",\\\"message\\\":\\\"(0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:20:19.685746 6517 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:20:19.685774 6517 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:20:19.685815 6517 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:20:19.685850 6517 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:20:19.686345 6517 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:20:19.686442 6517 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:20:19.686483 6517 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:20:19.686637 6517 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 14:20:19.686649 6517 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 14:20:19.686672 6517 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 14:20:19.686694 6517 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 14:20:19.686719 6517 factory.go:656] Stopping watch factory\\\\nI1124 14:20:19.686740 6517 ovnkube.go:599] Stopped ovnkube\\\\nI1124 14:20:19.686758 6517 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 14:20:19.686812 6517 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 14:20:19.686970 6517 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.770859 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.787069 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5484ae55-c0f7-4186-ad8b-cf893035524b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d272d832cbfb5ccdd258a4151ebf2e0324964440edd34bd7e6751f9eff958e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.805262 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.825881 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.832914 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.832981 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.832993 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.833036 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.833049 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.843774 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.856406 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.866362 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.878816 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.896445 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.911060 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.929318 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.935127 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.935155 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.935164 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.935180 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.935190 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:43Z","lastTransitionTime":"2025-11-24T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.944288 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.959401 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:34Z\\\",\\\"message\\\":\\\"2025-11-24T14:19:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6\\\\n2025-11-24T14:19:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6 to /host/opt/cni/bin/\\\\n2025-11-24T14:19:49Z [verbose] multus-daemon started\\\\n2025-11-24T14:19:49Z [verbose] Readiness Indicator file check\\\\n2025-11-24T14:20:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.982239 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9466a71a-baf5-418d-82d1-ed9122b5b9ea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca06cc6cb37db7454f12a957471e578968131669b24119e2de9748230dcaa3b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6073ae0db844c158057de597185a6f429e235c6442988a6885b208be5f55560f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcdf6e8f860ef2a3419b49f53e2ff550d704467c862b5ce00d3946eb284fb284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f60e1b6e031acfddc99eaab21f25b1cf791148cb732891e9c251d2577c064a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54579d92e01f8b881c6600bc68e65c8c3afe1a52e6e14250077390ea376fcde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:43 crc kubenswrapper[4822]: I1124 14:20:43.995363 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:43Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.005491 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:44Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.016456 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:44Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.037795 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.037834 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.037845 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.037864 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.037877 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.140903 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.140970 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.140990 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.141020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.141041 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.244281 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.244367 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.244391 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.244421 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.244446 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.347176 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.347273 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.347295 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.347321 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.347340 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.450476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.450698 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.450708 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.450723 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.450734 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.553179 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.553261 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.553279 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.553305 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.553324 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.656618 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.656686 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.656712 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.656742 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.656762 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.760187 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.760323 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.760335 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.760352 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.760363 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.863468 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.863522 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.863539 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.863565 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.863583 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.967461 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.967546 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.967568 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.967603 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:44 crc kubenswrapper[4822]: I1124 14:20:44.967625 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:44Z","lastTransitionTime":"2025-11-24T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.070225 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.070270 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.070278 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.070293 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.070304 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.172457 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.172518 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.172534 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.172556 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.172576 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.275456 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.275513 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.275528 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.275549 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.275565 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.378578 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.378628 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.378643 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.378667 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.378684 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.481750 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.481832 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.481851 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.481877 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.481903 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.585331 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.585394 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.585402 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.585422 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.585433 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.687558 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.687607 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.687618 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.687634 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.687646 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.704047 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.704100 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:45 crc kubenswrapper[4822]: E1124 14:20:45.704162 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.704197 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.704249 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:45 crc kubenswrapper[4822]: E1124 14:20:45.704392 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:45 crc kubenswrapper[4822]: E1124 14:20:45.704543 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:45 crc kubenswrapper[4822]: E1124 14:20:45.704708 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.705426 4822 scope.go:117] "RemoveContainer" containerID="9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.791326 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.791397 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.791417 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.791448 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.791469 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.894162 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.894284 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.894318 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.894346 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.894378 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.998082 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.998118 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.998127 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.998141 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:45 crc kubenswrapper[4822]: I1124 14:20:45.998150 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:45Z","lastTransitionTime":"2025-11-24T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.101678 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.101727 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.101745 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.101767 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.101784 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:46Z","lastTransitionTime":"2025-11-24T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.203930 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.203973 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.203988 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.204009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.204026 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:46Z","lastTransitionTime":"2025-11-24T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.261152 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/2.log" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.264509 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.265256 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.279546 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.296793 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.306562 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.306612 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.306621 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.306637 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.306647 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:46Z","lastTransitionTime":"2025-11-24T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.312839 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.330062 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.378014 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.401852 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:34Z\\\",\\\"message\\\":\\\"2025-11-24T14:19:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6\\\\n2025-11-24T14:19:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6 to /host/opt/cni/bin/\\\\n2025-11-24T14:19:49Z [verbose] multus-daemon started\\\\n2025-11-24T14:19:49Z [verbose] Readiness Indicator file check\\\\n2025-11-24T14:20:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.409259 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.409784 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.410135 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.410163 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.410176 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:46Z","lastTransitionTime":"2025-11-24T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.426488 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9466a71a-baf5-418d-82d1-ed9122b5b9ea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca06cc6cb37db7454f12a957471e578968131669b24119e2de9748230dcaa3b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6073ae0db844c158057de597185a6f429e235c6442988a6885b208be5f55560f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcdf6e8f860ef2a3419b49f53e2ff550d704467c862b5ce00d3946eb284fb284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f60e1b6e031acfddc99eaab21f25b1cf791148cb732891e9c251d2577c064a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54579d92e01f8b881c6600bc68e65c8c3afe1a52e6e14250077390ea376fcde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.439947 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.450294 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.459823 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.481087 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:19Z\\\",\\\"message\\\":\\\"(0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:20:19.685746 6517 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:20:19.685774 6517 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:20:19.685815 6517 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:20:19.685850 6517 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:20:19.686345 6517 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:20:19.686442 6517 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:20:19.686483 6517 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:20:19.686637 6517 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 14:20:19.686649 6517 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 14:20:19.686672 6517 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 14:20:19.686694 6517 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 14:20:19.686719 6517 factory.go:656] Stopping watch factory\\\\nI1124 14:20:19.686740 6517 ovnkube.go:599] Stopped ovnkube\\\\nI1124 14:20:19.686758 6517 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 14:20:19.686812 6517 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 14:20:19.686970 6517 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.493630 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.503433 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5484ae55-c0f7-4186-ad8b-cf893035524b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d272d832cbfb5ccdd258a4151ebf2e0324964440edd34bd7e6751f9eff958e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.512965 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.513011 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.513020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.513036 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.513046 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:46Z","lastTransitionTime":"2025-11-24T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.516315 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.528683 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.540559 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.550855 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.575528 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.592500 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:46Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.615419 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.615481 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.615508 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.615528 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.615541 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:46Z","lastTransitionTime":"2025-11-24T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.719646 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.719713 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.719737 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.719772 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.719797 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:46Z","lastTransitionTime":"2025-11-24T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.829377 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.829454 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.829467 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.829485 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.829497 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:46Z","lastTransitionTime":"2025-11-24T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.932405 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.932474 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.932484 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.932524 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:46 crc kubenswrapper[4822]: I1124 14:20:46.932534 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:46Z","lastTransitionTime":"2025-11-24T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.035071 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.035430 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.035444 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.035461 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.035473 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.138457 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.138492 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.138502 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.138520 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.138532 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.241034 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.241080 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.241094 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.241110 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.241121 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.270929 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/3.log" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.271720 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/2.log" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.276161 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa" exitCode=1 Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.276221 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.276261 4822 scope.go:117] "RemoveContainer" containerID="9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.277057 4822 scope.go:117] "RemoveContainer" containerID="3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa" Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.277302 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.302742 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.323661 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.343861 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.343925 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.343940 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.343962 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.343982 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.345350 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:34Z\\\",\\\"message\\\":\\\"2025-11-24T14:19:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6\\\\n2025-11-24T14:19:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6 to /host/opt/cni/bin/\\\\n2025-11-24T14:19:49Z [verbose] multus-daemon started\\\\n2025-11-24T14:19:49Z [verbose] Readiness Indicator file check\\\\n2025-11-24T14:20:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.371431 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9466a71a-baf5-418d-82d1-ed9122b5b9ea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca06cc6cb37db7454f12a957471e578968131669b24119e2de9748230dcaa3b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6073ae0db844c158057de597185a6f429e235c6442988a6885b208be5f55560f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcdf6e8f860ef2a3419b49f53e2ff550d704467c862b5ce00d3946eb284fb284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f60e1b6e031acfddc99eaab21f25b1cf791148cb732891e9c251d2577c064a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54579d92e01f8b881c6600bc68e65c8c3afe1a52e6e14250077390ea376fcde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.388960 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.400852 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.410963 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.431192 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.447721 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.447788 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.447812 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.447840 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.447858 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.464459 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9478ef2bc5e5256eedef1788ab0af74ecc0153eb075c93ddcfce16d4056a01a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:19Z\\\",\\\"message\\\":\\\"(0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 14:20:19.685746 6517 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 14:20:19.685774 6517 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 14:20:19.685815 6517 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 14:20:19.685850 6517 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 14:20:19.686345 6517 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 14:20:19.686442 6517 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 14:20:19.686483 6517 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 14:20:19.686637 6517 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 14:20:19.686649 6517 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 14:20:19.686672 6517 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 14:20:19.686694 6517 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 14:20:19.686719 6517 factory.go:656] Stopping watch factory\\\\nI1124 14:20:19.686740 6517 ovnkube.go:599] Stopped ovnkube\\\\nI1124 14:20:19.686758 6517 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 14:20:19.686812 6517 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 14:20:19.686970 6517 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:46Z\\\",\\\"message\\\":\\\" Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 14:20:46.654624 6884 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-nst99\\\\nI1124 14:20:46.654772 6884 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-nst99 in node crc\\\\nI1124 14:20:46.654787 6884 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-nst99 after 0 failed attempt(s)\\\\nI1124 14:20:46.654795 6884 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-nst99\\\\nI1124 14:20:46.654673 6884 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nF1124 14:20:46.654807 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.487923 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.501962 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5484ae55-c0f7-4186-ad8b-cf893035524b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d272d832cbfb5ccdd258a4151ebf2e0324964440edd34bd7e6751f9eff958e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.519811 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.538427 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.550614 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.550688 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.550708 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.550731 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.550749 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.559392 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.580655 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.597500 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.597710 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.597770 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.597816 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.597863 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.597978 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:51.597940033 +0000 UTC m=+148.714580550 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598079 4822 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598163 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598226 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598229 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:21:51.59818554 +0000 UTC m=+148.714826017 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598243 4822 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598105 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598312 4822 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598323 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 14:21:51.598297324 +0000 UTC m=+148.714938021 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598345 4822 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598104 4822 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598392 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 14:21:51.598380167 +0000 UTC m=+148.715020874 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.598427 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 14:21:51.598404657 +0000 UTC m=+148.715045164 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.600719 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.621837 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.640543 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.653645 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.653707 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.653724 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.653748 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.653766 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.657656 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:47Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.703656 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.703734 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.703757 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.704138 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.704117 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.703887 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.704336 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:47 crc kubenswrapper[4822]: E1124 14:20:47.704455 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.756903 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.756966 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.756985 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.757009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.757026 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.860091 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.860423 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.860615 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.860774 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.860901 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.964824 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.964895 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.964917 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.964947 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:47 crc kubenswrapper[4822]: I1124 14:20:47.964968 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:47Z","lastTransitionTime":"2025-11-24T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.067968 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.068059 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.068085 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.068121 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.068148 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:48Z","lastTransitionTime":"2025-11-24T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.171991 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.172053 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.172070 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.172100 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.172121 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:48Z","lastTransitionTime":"2025-11-24T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.275424 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.275772 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.275903 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.276038 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.276167 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:48Z","lastTransitionTime":"2025-11-24T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.283294 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/3.log" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.288425 4822 scope.go:117] "RemoveContainer" containerID="3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa" Nov 24 14:20:48 crc kubenswrapper[4822]: E1124 14:20:48.288829 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.313686 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.335975 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.350600 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:34Z\\\",\\\"message\\\":\\\"2025-11-24T14:19:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6\\\\n2025-11-24T14:19:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6 to /host/opt/cni/bin/\\\\n2025-11-24T14:19:49Z [verbose] multus-daemon started\\\\n2025-11-24T14:19:49Z [verbose] Readiness Indicator file check\\\\n2025-11-24T14:20:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.365627 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.379076 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.379123 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.379133 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.379148 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.379158 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:48Z","lastTransitionTime":"2025-11-24T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.391026 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9466a71a-baf5-418d-82d1-ed9122b5b9ea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca06cc6cb37db7454f12a957471e578968131669b24119e2de9748230dcaa3b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6073ae0db844c158057de597185a6f429e235c6442988a6885b208be5f55560f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcdf6e8f860ef2a3419b49f53e2ff550d704467c862b5ce00d3946eb284fb284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f60e1b6e031acfddc99eaab21f25b1cf791148cb732891e9c251d2577c064a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54579d92e01f8b881c6600bc68e65c8c3afe1a52e6e14250077390ea376fcde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.411371 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.425294 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.446431 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.460821 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.481906 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.482086 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.482458 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.482479 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.482500 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.482519 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:48Z","lastTransitionTime":"2025-11-24T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.505718 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:46Z\\\",\\\"message\\\":\\\" Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 14:20:46.654624 6884 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-nst99\\\\nI1124 14:20:46.654772 6884 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-nst99 in node crc\\\\nI1124 14:20:46.654787 6884 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-nst99 after 0 failed attempt(s)\\\\nI1124 14:20:46.654795 6884 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-nst99\\\\nI1124 14:20:46.654673 6884 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nF1124 14:20:46.654807 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.533116 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.547345 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5484ae55-c0f7-4186-ad8b-cf893035524b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d272d832cbfb5ccdd258a4151ebf2e0324964440edd34bd7e6751f9eff958e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.563473 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.579315 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.585728 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.585787 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.585804 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.585829 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.585845 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:48Z","lastTransitionTime":"2025-11-24T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.597263 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.614867 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.632712 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.648373 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:48Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.688955 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.689023 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.689049 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.689079 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.689103 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:48Z","lastTransitionTime":"2025-11-24T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.791982 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.792055 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.792078 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.792104 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.792124 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:48Z","lastTransitionTime":"2025-11-24T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.900715 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.900840 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.900870 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.900908 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:48 crc kubenswrapper[4822]: I1124 14:20:48.900936 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:48Z","lastTransitionTime":"2025-11-24T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.004271 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.004335 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.004353 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.004377 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.004395 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.106826 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.106902 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.106924 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.106953 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.106975 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.210193 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.210291 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.210308 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.210334 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.210353 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.314363 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.314437 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.314460 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.314492 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.314517 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.417198 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.417256 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.417267 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.417285 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.417296 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.520329 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.520406 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.520427 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.520452 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.520471 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.624766 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.624854 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.624877 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.624904 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.624924 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.704011 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.704106 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.704129 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:49 crc kubenswrapper[4822]: E1124 14:20:49.704314 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.704360 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:49 crc kubenswrapper[4822]: E1124 14:20:49.704494 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:49 crc kubenswrapper[4822]: E1124 14:20:49.704601 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:49 crc kubenswrapper[4822]: E1124 14:20:49.704714 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.727484 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.727538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.727555 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.727581 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.727603 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.831316 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.831456 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.831479 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.831534 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.831556 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.934991 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.935058 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.935075 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.935098 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:49 crc kubenswrapper[4822]: I1124 14:20:49.935113 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:49Z","lastTransitionTime":"2025-11-24T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.038332 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.038404 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.038424 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.038450 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.038469 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.141403 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.141496 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.141521 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.141695 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.141724 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.244984 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.245064 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.245087 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.245114 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.245132 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.349045 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.349142 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.349168 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.349202 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.349266 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.452700 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.452765 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.452782 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.452808 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.452828 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.555286 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.555360 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.555384 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.555414 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.555431 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.658328 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.658395 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.658412 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.658438 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.658455 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.761070 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.761275 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.761305 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.761383 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.761404 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.864800 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.864938 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.864962 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.864995 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.865016 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.968878 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.968948 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.968974 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.969007 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:50 crc kubenswrapper[4822]: I1124 14:20:50.969029 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:50Z","lastTransitionTime":"2025-11-24T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.072871 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.072934 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.072951 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.072975 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.072992 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.149925 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.149997 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.150015 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.150040 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.150066 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.174395 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.182132 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.182249 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.182275 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.182304 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.182329 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.209388 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.216497 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.216536 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.216548 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.216565 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.216577 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.238429 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.246037 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.246104 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.246124 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.246151 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.246180 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.271644 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.276401 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.276457 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.276473 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.276498 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.276515 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.297122 4822 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"498687cf-1a90-4e07-8490-79ba0b3ff4d4\\\",\\\"systemUUID\\\":\\\"4bde0dbd-3112-410e-b715-484aff4b0fa5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:51Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.297268 4822 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.299690 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.299750 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.299770 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.299799 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.299821 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.402946 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.403009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.403034 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.403065 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.403088 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.506045 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.506091 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.506110 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.506134 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.506152 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.609193 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.609346 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.609366 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.609397 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.609420 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.704401 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.704455 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.704492 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.704632 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.704958 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.705083 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.705472 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:51 crc kubenswrapper[4822]: E1124 14:20:51.705739 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.712426 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.712494 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.712519 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.712549 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.712573 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.815474 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.815546 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.815566 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.815592 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.815611 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.921831 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.921880 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.921897 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.921921 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:51 crc kubenswrapper[4822]: I1124 14:20:51.921939 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:51Z","lastTransitionTime":"2025-11-24T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.025330 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.025398 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.025415 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.025446 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.025466 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.128842 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.128913 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.128939 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.128969 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.128994 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.232380 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.232432 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.232444 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.232462 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.232479 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.336337 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.336398 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.336418 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.336445 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.336463 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.439564 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.439645 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.439668 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.439702 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.439726 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.543476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.543544 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.543565 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.543593 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.543611 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.648018 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.648070 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.648115 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.648135 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.648151 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.751327 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.751389 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.751405 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.751432 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.751450 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.855163 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.855259 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.855283 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.855316 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.855337 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.958297 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.958360 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.958387 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.958416 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:52 crc kubenswrapper[4822]: I1124 14:20:52.958441 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:52Z","lastTransitionTime":"2025-11-24T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.062145 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.062279 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.062297 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.062331 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.062349 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.164892 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.164963 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.164980 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.165006 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.165025 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.269379 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.269465 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.269489 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.269519 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.269543 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.377932 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.377985 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.377997 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.378019 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.378259 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.482689 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.482753 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.482815 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.482870 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.482891 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.585851 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.585913 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.585930 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.585954 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.585974 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.689503 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.689569 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.689587 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.689614 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.689634 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.703445 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:53 crc kubenswrapper[4822]: E1124 14:20:53.703780 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.703576 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.703541 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.703587 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:53 crc kubenswrapper[4822]: E1124 14:20:53.704538 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:53 crc kubenswrapper[4822]: E1124 14:20:53.704657 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:53 crc kubenswrapper[4822]: E1124 14:20:53.704672 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.722359 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb351a60-3e17-4d81-815c-39bf874cf341\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20b5b96e5041f4c54ccbfc8f5b265dff51ac231bb7696f15227f8c22e20ba44b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50e9fdeeec603e0ef50d5e9dab00faeb27cda714eba5c6cb0eac682861a0609c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz925\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wfbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.739529 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5484ae55-c0f7-4186-ad8b-cf893035524b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d272d832cbfb5ccdd258a4151ebf2e0324964440edd34bd7e6751f9eff958e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfd8135dd1b668312f90b4b41211df5f8a329fa3934bff57fd539caa0b0a2877\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.755795 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9264a29-95eb-4f97-9f42-44f2c23b3cb2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac2913f06ad696a463745d786010d39634a18ab00cbf58ac6db626d5d19504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bac90c681133091738398f94b0e0763026d7efaae0e9c9596449948d7c4cbbfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f8229c5c7e987c7f83d1523dd0cb896b36e0ae78ff5e5acf672b0d417d12062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7647af53154dbce49b359a83661cf4921655cd67fd36b124102a8eb7528de322\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.772984 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.790427 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce21009e6d29e94376b2f422ebdddabb470ffbab943b5f90b1eb2cdcfd0f4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948eb65db1323e515ba76f5863890c54b1b517cd72ae8b9dfa70e97c01916fe6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.791983 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.792064 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.792086 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.792113 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.792134 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.806244 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46d4dd42386ba50722a53652a520e3dcbb197575d4df6ac766d011b5964542d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.827289 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45082ac5-a419-4ce2-a11b-e9a6b8178698\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e05389dea3fd294be1c5e2e0bdf152393562f47c08310803463a0879c9443ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfab82292c58dcfbb16f70091cad14dc5517977ed13a76c49ff5e7f00f95eeae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9fde42a6933533074906b740c2fd71a4b3e26b998bfbc7fd09004655102d89f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7aeb53788682cb8af222c7768db960106e81c50883b20df3b57443449cd48249\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46d0da27b540cce36d7941c5de1c64a745328b1f3828464acd6272900940cb87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8d64be8fbd220a43dcb2a2699c2ee863fb049c3c03d2fe9f2d45f95c26d2219\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d111f72de42b32f1ef143d14dc2b6298eb26134480e9cb79ad3355132ec682\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8dsg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqrzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.855764 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13d10aa8-fab6-40df-90c8-0784ebac8e0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:46Z\\\",\\\"message\\\":\\\" Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 14:20:46.654624 6884 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-nst99\\\\nI1124 14:20:46.654772 6884 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-nst99 in node crc\\\\nI1124 14:20:46.654787 6884 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-nst99 after 0 failed attempt(s)\\\\nI1124 14:20:46.654795 6884 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-nst99\\\\nI1124 14:20:46.654673 6884 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nF1124 14:20:46.654807 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:20:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhtnb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-swpcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.874263 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dljmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5prnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:20:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dljmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.893302 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.897075 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.897148 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.897166 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.897192 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.897239 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.912605 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.932611 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cb8d340-5782-4aa4-b81e-e1b64e6b5207\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a892d31f8f91b8f1af88e53869bfae790009f2c4ede9f834677cc926b6fedd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkh7k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nst99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.953553 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71025cbc-9d01-40eb-9ef0-e81095c47240\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20e4b2546eefdf007801713508d8b39315ff01b6b10462648cf0c76888dadcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fc234734593a63fef5d1c2556185a727fa0fb840db0879c469723afb29fa1de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fb48887c0ca2edc2ab241332d41cb690e834dd4b9c5536664be75efe3f5f378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d8a6a53d1420e923baceff136d56a57bf07b37629ba6cf4729fe96d25aeb57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88465b012b0ab530503798dd703c7d3064d448743975195152fa9ddfaf0af75c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T14:19:37Z\\\",\\\"message\\\":\\\"W1124 14:19:26.845975 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 14:19:26.846285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763993966 cert, and key in /tmp/serving-cert-102943314/serving-signer.crt, /tmp/serving-cert-102943314/serving-signer.key\\\\nI1124 14:19:27.330047 1 observer_polling.go:159] Starting file observer\\\\nW1124 14:19:27.334470 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 14:19:27.334698 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 14:19:27.337439 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-102943314/tls.crt::/tmp/serving-cert-102943314/tls.key\\\\\\\"\\\\nF1124 14:19:37.696869 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea43672c471796a6d8c229390df9870d52fb2fd041d4fbd8defa5f31b1e5215d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba21c32d520a285269be7bf3858129a71595079b975341f816bedf83c5c1197\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.972501 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e49fa87-132b-4ae9-9f64-46cc07b9951f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b1b95daae774bc4722d5e434272f1625a287c5c722dac05e191fcaf149de1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da4e4455fab7ad80752ad3bbca3a0e0d5f28534e2817c8454360ce198910ec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee90115274f8951cf888e19291b4b4fe60f6637ae26922b91043c6f822296318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8cd991b05c02bbebe6b670e7eef2a0b7b8c199288bf288ac6b30f2731177de7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.993064 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-d6rbt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92c230aa-aed2-4b07-82bd-98b1d2ffa456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:20:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T14:20:34Z\\\",\\\"message\\\":\\\"2025-11-24T14:19:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6\\\\n2025-11-24T14:19:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3c446be7-5460-46fe-9a4d-7d356720f2e6 to /host/opt/cni/bin/\\\\n2025-11-24T14:19:49Z [verbose] multus-daemon started\\\\n2025-11-24T14:19:49Z [verbose] Readiness Indicator file check\\\\n2025-11-24T14:20:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:20:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8qtc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-d6rbt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:53Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.999760 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:53 crc kubenswrapper[4822]: I1124 14:20:53.999818 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:53.999830 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:53.999849 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:53.999867 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:53Z","lastTransitionTime":"2025-11-24T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.026419 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9466a71a-baf5-418d-82d1-ed9122b5b9ea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca06cc6cb37db7454f12a957471e578968131669b24119e2de9748230dcaa3b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6073ae0db844c158057de597185a6f429e235c6442988a6885b208be5f55560f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcdf6e8f860ef2a3419b49f53e2ff550d704467c862b5ce00d3946eb284fb284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f60e1b6e031acfddc99eaab21f25b1cf791148cb732891e9c251d2577c064a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54579d92e01f8b881c6600bc68e65c8c3afe1a52e6e14250077390ea376fcde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adb5ed048c76df300f996179488bd14f0325369c3f7f4eb2f9c0feafd3f8b3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0565035972b8f668201228adf5e59d9fd75d5fe8575cf74294ee5a19c8e668f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:25Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d63f0c259663f0d960f691bd07356e01b3d39577906d2aa9ee683ab4b5c51527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T14:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T14:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.047916 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3557bfe4e6efce26e3f656e7d2a277540813ad28b2557ba64e2d74dc5b806564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.065174 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kng5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f878d17-84e9-40c4-b161-eab718c34373\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c83f6e984375d365146152bb7d3157df01c79cdc7e6f2169b9a2a6178bb71a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4l2l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kng5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.081393 4822 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x8rxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05c7955-ab58-4997-847d-559aa181ea8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T14:19:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc03307716efc1f6e9021aabfb929c2fc12293936da538563171e1be590fd1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T14:19:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r7ckq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T14:19:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x8rxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T14:20:54Z is after 2025-08-24T17:21:41Z" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.103072 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.103131 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.103150 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.103175 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.103194 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:54Z","lastTransitionTime":"2025-11-24T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.206808 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.206867 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.206891 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.206917 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.206937 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:54Z","lastTransitionTime":"2025-11-24T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.311240 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.311301 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.311319 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.311343 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.311365 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:54Z","lastTransitionTime":"2025-11-24T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.414295 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.414365 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.414383 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.414409 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.414429 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:54Z","lastTransitionTime":"2025-11-24T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.518603 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.518681 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.518700 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.518725 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.518742 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:54Z","lastTransitionTime":"2025-11-24T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.622678 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.622749 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.622767 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.622797 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.622816 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:54Z","lastTransitionTime":"2025-11-24T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.725502 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.725596 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.725615 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.725642 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.725661 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:54Z","lastTransitionTime":"2025-11-24T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.828277 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.828396 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.828422 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.828455 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.828482 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:54Z","lastTransitionTime":"2025-11-24T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.930972 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.931011 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.931023 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.931040 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:54 crc kubenswrapper[4822]: I1124 14:20:54.931052 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:54Z","lastTransitionTime":"2025-11-24T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.034480 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.034517 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.034533 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.034568 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.034581 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.136674 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.136717 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.136730 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.136745 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.136755 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.239358 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.239475 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.239493 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.239514 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.239530 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.341591 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.341641 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.341653 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.341670 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.341682 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.444494 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.444544 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.444552 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.444566 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.444581 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.547498 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.547588 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.547605 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.547632 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.547651 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.650909 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.650983 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.651006 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.651039 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.651064 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.704089 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.704196 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.704343 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.704379 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:55 crc kubenswrapper[4822]: E1124 14:20:55.704343 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:55 crc kubenswrapper[4822]: E1124 14:20:55.704770 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:55 crc kubenswrapper[4822]: E1124 14:20:55.705107 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:55 crc kubenswrapper[4822]: E1124 14:20:55.705506 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.753557 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.753595 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.753606 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.753638 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.753649 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.856820 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.856868 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.856879 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.856897 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.856909 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.960275 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.960363 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.960397 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.960465 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:55 crc kubenswrapper[4822]: I1124 14:20:55.960489 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:55Z","lastTransitionTime":"2025-11-24T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.063852 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.063910 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.063931 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.063959 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.063983 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.166948 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.167020 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.167038 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.167062 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.167080 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.270544 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.270620 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.270645 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.270679 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.270706 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.373153 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.373241 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.373253 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.373272 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.373284 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.475899 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.475980 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.476004 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.476034 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.476055 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.578954 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.578988 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.578996 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.579009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.579018 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.681876 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.681934 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.681946 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.681965 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.681980 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.785327 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.785398 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.785423 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.785465 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.785483 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.888621 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.888662 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.888677 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.888700 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.888716 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.991879 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.991927 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.991936 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.991952 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:56 crc kubenswrapper[4822]: I1124 14:20:56.991965 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:56Z","lastTransitionTime":"2025-11-24T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.095324 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.095403 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.095421 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.095452 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.095472 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:57Z","lastTransitionTime":"2025-11-24T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.198476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.198529 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.198542 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.198560 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.198573 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:57Z","lastTransitionTime":"2025-11-24T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.301440 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.301541 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.301560 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.301614 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.301636 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:57Z","lastTransitionTime":"2025-11-24T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.404365 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.404447 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.404464 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.404513 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.404581 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:57Z","lastTransitionTime":"2025-11-24T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.508474 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.508518 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.508558 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.508574 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.508584 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:57Z","lastTransitionTime":"2025-11-24T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.611875 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.611960 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.611982 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.612012 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.612034 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:57Z","lastTransitionTime":"2025-11-24T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.704180 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.704370 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.704370 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:57 crc kubenswrapper[4822]: E1124 14:20:57.704573 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.704615 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:57 crc kubenswrapper[4822]: E1124 14:20:57.704742 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:57 crc kubenswrapper[4822]: E1124 14:20:57.704845 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:57 crc kubenswrapper[4822]: E1124 14:20:57.704942 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.714611 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.714658 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.714678 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.714701 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.714719 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:57Z","lastTransitionTime":"2025-11-24T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.818911 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.818967 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.818983 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.819009 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.819026 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:57Z","lastTransitionTime":"2025-11-24T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.922194 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.922254 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.922266 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.922284 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:57 crc kubenswrapper[4822]: I1124 14:20:57.922298 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:57Z","lastTransitionTime":"2025-11-24T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.025538 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.025622 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.025646 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.025676 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.025699 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.128315 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.128391 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.128414 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.128447 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.128471 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.231649 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.231710 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.231721 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.231739 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.231751 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.334769 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.334832 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.334852 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.334875 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.334896 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.438013 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.438059 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.438073 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.438094 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.438110 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.541608 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.541701 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.541740 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.541776 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.541807 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.645455 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.645520 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.645530 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.645547 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.645558 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.748463 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.748543 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.748569 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.748639 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.748661 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.850587 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.850650 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.850660 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.850677 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.850687 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.954366 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.954435 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.954454 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.954483 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:58 crc kubenswrapper[4822]: I1124 14:20:58.954515 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:58Z","lastTransitionTime":"2025-11-24T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.057798 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.057861 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.057877 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.057901 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.057919 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.160497 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.160582 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.160637 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.160671 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.160694 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.263796 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.263872 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.263891 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.263917 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.263938 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.367722 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.367784 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.367800 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.367852 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.367870 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.470939 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.471016 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.471035 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.471062 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.471081 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.574618 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.574677 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.574697 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.574721 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.574740 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.678161 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.678281 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.678308 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.678340 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.678363 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.703830 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.703918 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.703838 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:20:59 crc kubenswrapper[4822]: E1124 14:20:59.704038 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.704316 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:20:59 crc kubenswrapper[4822]: E1124 14:20:59.704955 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:20:59 crc kubenswrapper[4822]: E1124 14:20:59.705159 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.705470 4822 scope.go:117] "RemoveContainer" containerID="3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa" Nov 24 14:20:59 crc kubenswrapper[4822]: E1124 14:20:59.705452 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:20:59 crc kubenswrapper[4822]: E1124 14:20:59.705814 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.780816 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.780861 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.780871 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.780889 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.780904 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.884534 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.884598 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.884613 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.884661 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.884674 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.987820 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.987942 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.987960 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.988029 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:20:59 crc kubenswrapper[4822]: I1124 14:20:59.988047 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:20:59Z","lastTransitionTime":"2025-11-24T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.091632 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.091716 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.091740 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.091776 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.091834 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:00Z","lastTransitionTime":"2025-11-24T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.195469 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.195519 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.195537 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.195562 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.195579 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:00Z","lastTransitionTime":"2025-11-24T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.299544 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.299619 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.299636 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.299659 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.299677 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:00Z","lastTransitionTime":"2025-11-24T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.403442 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.403526 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.403550 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.403573 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.403591 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:00Z","lastTransitionTime":"2025-11-24T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.506504 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.506582 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.506607 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.506636 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.506658 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:00Z","lastTransitionTime":"2025-11-24T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.609597 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.609667 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.609718 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.609751 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.609775 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:00Z","lastTransitionTime":"2025-11-24T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.712679 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.712752 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.712775 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.712798 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.712815 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:00Z","lastTransitionTime":"2025-11-24T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.816259 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.816336 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.816363 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.816393 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.816419 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:00Z","lastTransitionTime":"2025-11-24T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.919715 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.919787 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.919806 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.919832 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:00 crc kubenswrapper[4822]: I1124 14:21:00.919854 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:00Z","lastTransitionTime":"2025-11-24T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.023292 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.023349 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.023366 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.023391 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.023410 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:01Z","lastTransitionTime":"2025-11-24T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.126773 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.126848 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.126873 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.126902 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.126924 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:01Z","lastTransitionTime":"2025-11-24T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.230303 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.230386 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.230408 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.230441 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.230464 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:01Z","lastTransitionTime":"2025-11-24T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.333785 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.333852 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.333872 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.333900 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.333919 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:01Z","lastTransitionTime":"2025-11-24T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.425013 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.425071 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.425086 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.425110 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.425125 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:01Z","lastTransitionTime":"2025-11-24T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.452476 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.452544 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.452567 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.452599 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.452622 4822 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T14:21:01Z","lastTransitionTime":"2025-11-24T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.490146 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj"] Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.490798 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.494397 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.494695 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.495323 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.497429 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.534003 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=20.533963085 podStartE2EDuration="20.533963085s" podCreationTimestamp="2025-11-24 14:20:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.533966675 +0000 UTC m=+98.650607242" watchObservedRunningTime="2025-11-24 14:21:01.533963085 +0000 UTC m=+98.650603602" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.578815 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/264386c1-bc90-4d97-886d-f710dad65d74-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.578930 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/264386c1-bc90-4d97-886d-f710dad65d74-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.579090 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/264386c1-bc90-4d97-886d-f710dad65d74-service-ca\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.579144 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/264386c1-bc90-4d97-886d-f710dad65d74-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.579275 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/264386c1-bc90-4d97-886d-f710dad65d74-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.581252 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-kng5k" podStartSLOduration=75.581195917 podStartE2EDuration="1m15.581195917s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.580509075 +0000 UTC m=+98.697149592" watchObservedRunningTime="2025-11-24 14:21:01.581195917 +0000 UTC m=+98.697836434" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.597187 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-x8rxd" podStartSLOduration=75.597152818 podStartE2EDuration="1m15.597152818s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.596804377 +0000 UTC m=+98.713444904" watchObservedRunningTime="2025-11-24 14:21:01.597152818 +0000 UTC m=+98.713793335" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.660193 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-nqrzx" podStartSLOduration=75.660174637 podStartE2EDuration="1m15.660174637s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.627481849 +0000 UTC m=+98.744122406" watchObservedRunningTime="2025-11-24 14:21:01.660174637 +0000 UTC m=+98.776815114" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.672410 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wfbzd" podStartSLOduration=74.672392327 podStartE2EDuration="1m14.672392327s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.672022835 +0000 UTC m=+98.788663322" watchObservedRunningTime="2025-11-24 14:21:01.672392327 +0000 UTC m=+98.789032804" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.679852 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/264386c1-bc90-4d97-886d-f710dad65d74-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.679900 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/264386c1-bc90-4d97-886d-f710dad65d74-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.679951 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/264386c1-bc90-4d97-886d-f710dad65d74-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.679954 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/264386c1-bc90-4d97-886d-f710dad65d74-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.679981 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/264386c1-bc90-4d97-886d-f710dad65d74-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.680020 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/264386c1-bc90-4d97-886d-f710dad65d74-service-ca\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.680414 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/264386c1-bc90-4d97-886d-f710dad65d74-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.680880 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/264386c1-bc90-4d97-886d-f710dad65d74-service-ca\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.695304 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/264386c1-bc90-4d97-886d-f710dad65d74-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.702506 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/264386c1-bc90-4d97-886d-f710dad65d74-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-q87lj\" (UID: \"264386c1-bc90-4d97-886d-f710dad65d74\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.703534 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:01 crc kubenswrapper[4822]: E1124 14:21:01.703785 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.703589 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:01 crc kubenswrapper[4822]: E1124 14:21:01.704051 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.703548 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:01 crc kubenswrapper[4822]: E1124 14:21:01.704366 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.706420 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:01 crc kubenswrapper[4822]: E1124 14:21:01.706676 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.717086 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=24.717064718 podStartE2EDuration="24.717064718s" podCreationTimestamp="2025-11-24 14:20:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.697135719 +0000 UTC m=+98.813776206" watchObservedRunningTime="2025-11-24 14:21:01.717064718 +0000 UTC m=+98.833705195" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.728863 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.728839285 podStartE2EDuration="48.728839285s" podCreationTimestamp="2025-11-24 14:20:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.717542463 +0000 UTC m=+98.834182940" watchObservedRunningTime="2025-11-24 14:21:01.728839285 +0000 UTC m=+98.845479762" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.802755 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podStartSLOduration=75.802729051 podStartE2EDuration="1m15.802729051s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.802562946 +0000 UTC m=+98.919203433" watchObservedRunningTime="2025-11-24 14:21:01.802729051 +0000 UTC m=+98.919369538" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.812986 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.840949 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.840927874 podStartE2EDuration="1m18.840927874s" podCreationTimestamp="2025-11-24 14:19:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.821694548 +0000 UTC m=+98.938335055" watchObservedRunningTime="2025-11-24 14:21:01.840927874 +0000 UTC m=+98.957568361" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.841582 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.841576255 podStartE2EDuration="1m15.841576255s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.840268293 +0000 UTC m=+98.956908770" watchObservedRunningTime="2025-11-24 14:21:01.841576255 +0000 UTC m=+98.958216742" Nov 24 14:21:01 crc kubenswrapper[4822]: I1124 14:21:01.858752 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-d6rbt" podStartSLOduration=75.858723634 podStartE2EDuration="1m15.858723634s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:01.858640362 +0000 UTC m=+98.975280849" watchObservedRunningTime="2025-11-24 14:21:01.858723634 +0000 UTC m=+98.975364121" Nov 24 14:21:02 crc kubenswrapper[4822]: I1124 14:21:02.346527 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" event={"ID":"264386c1-bc90-4d97-886d-f710dad65d74","Type":"ContainerStarted","Data":"c673e9efa75d1411d60cdaf46c1fcb31a39c53108d72bb62ce9a17ad2799de01"} Nov 24 14:21:02 crc kubenswrapper[4822]: I1124 14:21:02.346588 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" event={"ID":"264386c1-bc90-4d97-886d-f710dad65d74","Type":"ContainerStarted","Data":"c14c70f8691700a7cb8ac039e0aec6f0a1519d50f603eb2e1a78667bf3602119"} Nov 24 14:21:02 crc kubenswrapper[4822]: I1124 14:21:02.363540 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q87lj" podStartSLOduration=76.363518158 podStartE2EDuration="1m16.363518158s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:02.363369203 +0000 UTC m=+99.480009760" watchObservedRunningTime="2025-11-24 14:21:02.363518158 +0000 UTC m=+99.480158625" Nov 24 14:21:03 crc kubenswrapper[4822]: I1124 14:21:03.703849 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:03 crc kubenswrapper[4822]: I1124 14:21:03.703969 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:03 crc kubenswrapper[4822]: I1124 14:21:03.704493 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:03 crc kubenswrapper[4822]: E1124 14:21:03.704494 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:03 crc kubenswrapper[4822]: I1124 14:21:03.704621 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:03 crc kubenswrapper[4822]: E1124 14:21:03.704742 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:03 crc kubenswrapper[4822]: E1124 14:21:03.704891 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:03 crc kubenswrapper[4822]: E1124 14:21:03.704985 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:05 crc kubenswrapper[4822]: I1124 14:21:05.532870 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:05 crc kubenswrapper[4822]: E1124 14:21:05.533064 4822 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:21:05 crc kubenswrapper[4822]: E1124 14:21:05.533173 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs podName:b8d0fb3c-6504-4242-b078-ee8c88e7f6f7 nodeName:}" failed. No retries permitted until 2025-11-24 14:22:09.533144905 +0000 UTC m=+166.649785432 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs") pod "network-metrics-daemon-dljmt" (UID: "b8d0fb3c-6504-4242-b078-ee8c88e7f6f7") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 14:21:05 crc kubenswrapper[4822]: I1124 14:21:05.703750 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:05 crc kubenswrapper[4822]: I1124 14:21:05.703809 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:05 crc kubenswrapper[4822]: I1124 14:21:05.703935 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:05 crc kubenswrapper[4822]: E1124 14:21:05.704141 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:05 crc kubenswrapper[4822]: I1124 14:21:05.704190 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:05 crc kubenswrapper[4822]: E1124 14:21:05.704395 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:05 crc kubenswrapper[4822]: E1124 14:21:05.704586 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:05 crc kubenswrapper[4822]: E1124 14:21:05.704764 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:07 crc kubenswrapper[4822]: I1124 14:21:07.703997 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:07 crc kubenswrapper[4822]: I1124 14:21:07.704076 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:07 crc kubenswrapper[4822]: I1124 14:21:07.704038 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:07 crc kubenswrapper[4822]: E1124 14:21:07.704292 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:07 crc kubenswrapper[4822]: I1124 14:21:07.704348 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:07 crc kubenswrapper[4822]: E1124 14:21:07.704565 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:07 crc kubenswrapper[4822]: E1124 14:21:07.704746 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:07 crc kubenswrapper[4822]: E1124 14:21:07.704943 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:09 crc kubenswrapper[4822]: I1124 14:21:09.704713 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:09 crc kubenswrapper[4822]: I1124 14:21:09.704900 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:09 crc kubenswrapper[4822]: I1124 14:21:09.705008 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:09 crc kubenswrapper[4822]: E1124 14:21:09.705186 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:09 crc kubenswrapper[4822]: I1124 14:21:09.705417 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:09 crc kubenswrapper[4822]: E1124 14:21:09.705394 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:09 crc kubenswrapper[4822]: E1124 14:21:09.705564 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:09 crc kubenswrapper[4822]: E1124 14:21:09.705679 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:11 crc kubenswrapper[4822]: I1124 14:21:11.703793 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:11 crc kubenswrapper[4822]: I1124 14:21:11.703826 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:11 crc kubenswrapper[4822]: I1124 14:21:11.703793 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:11 crc kubenswrapper[4822]: E1124 14:21:11.703995 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:11 crc kubenswrapper[4822]: I1124 14:21:11.703947 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:11 crc kubenswrapper[4822]: E1124 14:21:11.704234 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:11 crc kubenswrapper[4822]: E1124 14:21:11.704567 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:11 crc kubenswrapper[4822]: E1124 14:21:11.704665 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:12 crc kubenswrapper[4822]: I1124 14:21:12.704978 4822 scope.go:117] "RemoveContainer" containerID="3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa" Nov 24 14:21:12 crc kubenswrapper[4822]: E1124 14:21:12.705332 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:21:13 crc kubenswrapper[4822]: I1124 14:21:13.703582 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:13 crc kubenswrapper[4822]: I1124 14:21:13.703606 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:13 crc kubenswrapper[4822]: I1124 14:21:13.703606 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:13 crc kubenswrapper[4822]: E1124 14:21:13.705123 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:13 crc kubenswrapper[4822]: I1124 14:21:13.705266 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:13 crc kubenswrapper[4822]: E1124 14:21:13.705446 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:13 crc kubenswrapper[4822]: E1124 14:21:13.705590 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:13 crc kubenswrapper[4822]: E1124 14:21:13.705725 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:15 crc kubenswrapper[4822]: I1124 14:21:15.703835 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:15 crc kubenswrapper[4822]: I1124 14:21:15.703909 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:15 crc kubenswrapper[4822]: I1124 14:21:15.703837 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:15 crc kubenswrapper[4822]: I1124 14:21:15.704075 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:15 crc kubenswrapper[4822]: E1124 14:21:15.704067 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:15 crc kubenswrapper[4822]: E1124 14:21:15.704243 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:15 crc kubenswrapper[4822]: E1124 14:21:15.704376 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:15 crc kubenswrapper[4822]: E1124 14:21:15.704573 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:17 crc kubenswrapper[4822]: I1124 14:21:17.703626 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:17 crc kubenswrapper[4822]: I1124 14:21:17.704150 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:17 crc kubenswrapper[4822]: I1124 14:21:17.703719 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:17 crc kubenswrapper[4822]: E1124 14:21:17.704354 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:17 crc kubenswrapper[4822]: I1124 14:21:17.703643 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:17 crc kubenswrapper[4822]: E1124 14:21:17.704545 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:17 crc kubenswrapper[4822]: E1124 14:21:17.704645 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:17 crc kubenswrapper[4822]: E1124 14:21:17.704727 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:19 crc kubenswrapper[4822]: I1124 14:21:19.703516 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:19 crc kubenswrapper[4822]: I1124 14:21:19.703566 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:19 crc kubenswrapper[4822]: I1124 14:21:19.703681 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:19 crc kubenswrapper[4822]: E1124 14:21:19.703945 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:19 crc kubenswrapper[4822]: I1124 14:21:19.704061 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:19 crc kubenswrapper[4822]: E1124 14:21:19.704139 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:19 crc kubenswrapper[4822]: E1124 14:21:19.704248 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:19 crc kubenswrapper[4822]: E1124 14:21:19.704367 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.416666 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/1.log" Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.416973 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/0.log" Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.417008 4822 generic.go:334] "Generic (PLEG): container finished" podID="92c230aa-aed2-4b07-82bd-98b1d2ffa456" containerID="6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5" exitCode=1 Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.417035 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-d6rbt" event={"ID":"92c230aa-aed2-4b07-82bd-98b1d2ffa456","Type":"ContainerDied","Data":"6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5"} Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.417077 4822 scope.go:117] "RemoveContainer" containerID="385cb52762745610b26dfa582aa2c690c8b1a7b5117b1f22436b2c93653eaf16" Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.417850 4822 scope.go:117] "RemoveContainer" containerID="6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5" Nov 24 14:21:21 crc kubenswrapper[4822]: E1124 14:21:21.418182 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-d6rbt_openshift-multus(92c230aa-aed2-4b07-82bd-98b1d2ffa456)\"" pod="openshift-multus/multus-d6rbt" podUID="92c230aa-aed2-4b07-82bd-98b1d2ffa456" Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.703925 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.704007 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:21 crc kubenswrapper[4822]: E1124 14:21:21.704126 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.704431 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:21 crc kubenswrapper[4822]: I1124 14:21:21.704435 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:21 crc kubenswrapper[4822]: E1124 14:21:21.704583 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:21 crc kubenswrapper[4822]: E1124 14:21:21.704809 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:21 crc kubenswrapper[4822]: E1124 14:21:21.705320 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:22 crc kubenswrapper[4822]: I1124 14:21:22.421708 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/1.log" Nov 24 14:21:23 crc kubenswrapper[4822]: I1124 14:21:23.703908 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:23 crc kubenswrapper[4822]: I1124 14:21:23.704007 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:23 crc kubenswrapper[4822]: E1124 14:21:23.704043 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:23 crc kubenswrapper[4822]: I1124 14:21:23.704127 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:23 crc kubenswrapper[4822]: I1124 14:21:23.703926 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:23 crc kubenswrapper[4822]: E1124 14:21:23.704179 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:23 crc kubenswrapper[4822]: E1124 14:21:23.704397 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:23 crc kubenswrapper[4822]: E1124 14:21:23.704522 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:23 crc kubenswrapper[4822]: E1124 14:21:23.712179 4822 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 14:21:23 crc kubenswrapper[4822]: E1124 14:21:23.829681 4822 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:21:24 crc kubenswrapper[4822]: I1124 14:21:24.704694 4822 scope.go:117] "RemoveContainer" containerID="3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa" Nov 24 14:21:24 crc kubenswrapper[4822]: E1124 14:21:24.704903 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-swpcw_openshift-ovn-kubernetes(13d10aa8-fab6-40df-90c8-0784ebac8e0f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" Nov 24 14:21:25 crc kubenswrapper[4822]: I1124 14:21:25.704524 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:25 crc kubenswrapper[4822]: I1124 14:21:25.704664 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:25 crc kubenswrapper[4822]: I1124 14:21:25.704654 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:25 crc kubenswrapper[4822]: I1124 14:21:25.704775 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:25 crc kubenswrapper[4822]: E1124 14:21:25.704774 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:25 crc kubenswrapper[4822]: E1124 14:21:25.704940 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:25 crc kubenswrapper[4822]: E1124 14:21:25.705197 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:25 crc kubenswrapper[4822]: E1124 14:21:25.705297 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:27 crc kubenswrapper[4822]: I1124 14:21:27.703874 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:27 crc kubenswrapper[4822]: I1124 14:21:27.703970 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:27 crc kubenswrapper[4822]: E1124 14:21:27.704062 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:27 crc kubenswrapper[4822]: I1124 14:21:27.704080 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:27 crc kubenswrapper[4822]: I1124 14:21:27.704119 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:27 crc kubenswrapper[4822]: E1124 14:21:27.704211 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:27 crc kubenswrapper[4822]: E1124 14:21:27.704326 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:27 crc kubenswrapper[4822]: E1124 14:21:27.704623 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:28 crc kubenswrapper[4822]: E1124 14:21:28.831722 4822 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:21:29 crc kubenswrapper[4822]: I1124 14:21:29.704390 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:29 crc kubenswrapper[4822]: I1124 14:21:29.704491 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:29 crc kubenswrapper[4822]: I1124 14:21:29.704407 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:29 crc kubenswrapper[4822]: I1124 14:21:29.704433 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:29 crc kubenswrapper[4822]: E1124 14:21:29.704610 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:29 crc kubenswrapper[4822]: E1124 14:21:29.704811 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:29 crc kubenswrapper[4822]: E1124 14:21:29.704855 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:29 crc kubenswrapper[4822]: E1124 14:21:29.704931 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:31 crc kubenswrapper[4822]: I1124 14:21:31.703926 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:31 crc kubenswrapper[4822]: I1124 14:21:31.703976 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:31 crc kubenswrapper[4822]: E1124 14:21:31.704096 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:31 crc kubenswrapper[4822]: I1124 14:21:31.703958 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:31 crc kubenswrapper[4822]: I1124 14:21:31.704170 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:31 crc kubenswrapper[4822]: E1124 14:21:31.704296 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:31 crc kubenswrapper[4822]: E1124 14:21:31.704531 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:31 crc kubenswrapper[4822]: E1124 14:21:31.704646 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:33 crc kubenswrapper[4822]: I1124 14:21:33.703493 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:33 crc kubenswrapper[4822]: I1124 14:21:33.703448 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:33 crc kubenswrapper[4822]: I1124 14:21:33.703517 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:33 crc kubenswrapper[4822]: I1124 14:21:33.703566 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:33 crc kubenswrapper[4822]: E1124 14:21:33.705520 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:33 crc kubenswrapper[4822]: E1124 14:21:33.705994 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:33 crc kubenswrapper[4822]: E1124 14:21:33.706093 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:33 crc kubenswrapper[4822]: I1124 14:21:33.706291 4822 scope.go:117] "RemoveContainer" containerID="6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5" Nov 24 14:21:33 crc kubenswrapper[4822]: E1124 14:21:33.705797 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:33 crc kubenswrapper[4822]: E1124 14:21:33.832351 4822 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:21:34 crc kubenswrapper[4822]: I1124 14:21:34.469185 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/1.log" Nov 24 14:21:34 crc kubenswrapper[4822]: I1124 14:21:34.469309 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-d6rbt" event={"ID":"92c230aa-aed2-4b07-82bd-98b1d2ffa456","Type":"ContainerStarted","Data":"f528354f90b6adf0611b43f5ddea194a67abb6f33c1715df7b8dc36dab6a88f9"} Nov 24 14:21:35 crc kubenswrapper[4822]: I1124 14:21:35.704522 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:35 crc kubenswrapper[4822]: I1124 14:21:35.704584 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:35 crc kubenswrapper[4822]: E1124 14:21:35.704789 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:35 crc kubenswrapper[4822]: I1124 14:21:35.704845 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:35 crc kubenswrapper[4822]: I1124 14:21:35.704522 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:35 crc kubenswrapper[4822]: E1124 14:21:35.704962 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:35 crc kubenswrapper[4822]: E1124 14:21:35.705170 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:35 crc kubenswrapper[4822]: E1124 14:21:35.705331 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:36 crc kubenswrapper[4822]: I1124 14:21:36.705754 4822 scope.go:117] "RemoveContainer" containerID="3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa" Nov 24 14:21:37 crc kubenswrapper[4822]: I1124 14:21:37.481261 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/3.log" Nov 24 14:21:37 crc kubenswrapper[4822]: I1124 14:21:37.483942 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerStarted","Data":"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1"} Nov 24 14:21:37 crc kubenswrapper[4822]: I1124 14:21:37.484300 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:21:37 crc kubenswrapper[4822]: I1124 14:21:37.583526 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podStartSLOduration=111.583502225 podStartE2EDuration="1m51.583502225s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:37.516385196 +0000 UTC m=+134.633025693" watchObservedRunningTime="2025-11-24 14:21:37.583502225 +0000 UTC m=+134.700142722" Nov 24 14:21:37 crc kubenswrapper[4822]: I1124 14:21:37.584488 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-dljmt"] Nov 24 14:21:37 crc kubenswrapper[4822]: I1124 14:21:37.584634 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:37 crc kubenswrapper[4822]: E1124 14:21:37.584749 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:37 crc kubenswrapper[4822]: I1124 14:21:37.703793 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:37 crc kubenswrapper[4822]: I1124 14:21:37.703841 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:37 crc kubenswrapper[4822]: E1124 14:21:37.703959 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:37 crc kubenswrapper[4822]: E1124 14:21:37.704154 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:37 crc kubenswrapper[4822]: I1124 14:21:37.704372 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:37 crc kubenswrapper[4822]: E1124 14:21:37.704472 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:38 crc kubenswrapper[4822]: E1124 14:21:38.834604 4822 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:21:39 crc kubenswrapper[4822]: I1124 14:21:39.703796 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:39 crc kubenswrapper[4822]: I1124 14:21:39.703822 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:39 crc kubenswrapper[4822]: E1124 14:21:39.703984 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:39 crc kubenswrapper[4822]: I1124 14:21:39.704010 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:39 crc kubenswrapper[4822]: I1124 14:21:39.704063 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:39 crc kubenswrapper[4822]: E1124 14:21:39.704075 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:39 crc kubenswrapper[4822]: E1124 14:21:39.704171 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:39 crc kubenswrapper[4822]: E1124 14:21:39.704451 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:41 crc kubenswrapper[4822]: I1124 14:21:41.703827 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:41 crc kubenswrapper[4822]: I1124 14:21:41.703934 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:41 crc kubenswrapper[4822]: I1124 14:21:41.703949 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:41 crc kubenswrapper[4822]: I1124 14:21:41.704090 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:41 crc kubenswrapper[4822]: E1124 14:21:41.704073 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:41 crc kubenswrapper[4822]: E1124 14:21:41.704282 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:41 crc kubenswrapper[4822]: E1124 14:21:41.704526 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:41 crc kubenswrapper[4822]: E1124 14:21:41.704639 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:43 crc kubenswrapper[4822]: I1124 14:21:43.703604 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:43 crc kubenswrapper[4822]: I1124 14:21:43.703703 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:43 crc kubenswrapper[4822]: I1124 14:21:43.703604 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:43 crc kubenswrapper[4822]: I1124 14:21:43.703644 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:43 crc kubenswrapper[4822]: E1124 14:21:43.706262 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 14:21:43 crc kubenswrapper[4822]: E1124 14:21:43.706412 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dljmt" podUID="b8d0fb3c-6504-4242-b078-ee8c88e7f6f7" Nov 24 14:21:43 crc kubenswrapper[4822]: E1124 14:21:43.706570 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 14:21:43 crc kubenswrapper[4822]: E1124 14:21:43.706672 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.703867 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.703995 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.704077 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.704047 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.706376 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.706381 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.706376 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.706577 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.707680 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 14:21:45 crc kubenswrapper[4822]: I1124 14:21:45.713912 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.602602 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:51 crc kubenswrapper[4822]: E1124 14:21:51.602841 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:23:53.602801915 +0000 UTC m=+270.719442422 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.603097 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.603163 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.603256 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.603331 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.604776 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.611667 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.611955 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.612168 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.726316 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.736066 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 14:21:51 crc kubenswrapper[4822]: I1124 14:21:51.750755 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:52 crc kubenswrapper[4822]: W1124 14:21:52.223329 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-4438d061e91aca4fe61f8d04d0b40767c6366df15b16b19558f779448a0ab825 WatchSource:0}: Error finding container 4438d061e91aca4fe61f8d04d0b40767c6366df15b16b19558f779448a0ab825: Status 404 returned error can't find the container with id 4438d061e91aca4fe61f8d04d0b40767c6366df15b16b19558f779448a0ab825 Nov 24 14:21:52 crc kubenswrapper[4822]: W1124 14:21:52.230393 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-fd4b3490f7dcc1cec5454718623ae3ec78829bc647c090dd108340867bd90849 WatchSource:0}: Error finding container fd4b3490f7dcc1cec5454718623ae3ec78829bc647c090dd108340867bd90849: Status 404 returned error can't find the container with id fd4b3490f7dcc1cec5454718623ae3ec78829bc647c090dd108340867bd90849 Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.544821 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c0158e313c1afcb355ea8b8563090b6888327a3ac5b9b47b605da0d484e3e767"} Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.544900 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4438d061e91aca4fe61f8d04d0b40767c6366df15b16b19558f779448a0ab825"} Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.545109 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.546533 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1d9fca9f08cf9bd612bc1c6218836f17fb11c0902be628ad917ade8354b7d5c8"} Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.546593 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5b9347258f86c558608ed11562a823bdc3b2bc2aebe222a8c255e320f7fc219f"} Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.548837 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0ac1094cfcd1f166e0d10084bbbdd5205befbc73a7103ce37d1647472879f24c"} Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.548866 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fd4b3490f7dcc1cec5454718623ae3ec78829bc647c090dd108340867bd90849"} Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.831274 4822 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.881434 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qr4nb"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.882440 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.887070 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.889787 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.890517 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.890629 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-p9tw4"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.891657 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-p9tw4" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.892803 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.892930 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.895716 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ctks4"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.896642 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.897295 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.897485 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.899426 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.914265 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.914956 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gzxq"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.915373 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.915923 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.917718 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hl4bh"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.918307 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.918429 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.918796 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.919449 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.919824 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.919953 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.920249 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.920321 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.920525 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.920656 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.920748 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.920911 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.921057 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.921262 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.921511 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.922168 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.960448 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.960449 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.960779 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.963430 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.964953 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8bc7n"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.967988 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.968158 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.982112 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.982397 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.982629 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.982858 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.983985 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.984386 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.984679 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.984907 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5v7wp"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.985232 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.985592 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.986046 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.986699 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.986954 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.987562 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.987746 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.987847 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.987898 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.987983 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988062 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988095 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988193 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988346 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988558 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988673 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988766 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988774 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988848 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988885 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988977 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.988851 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.989448 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.989640 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.990083 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.990290 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.991872 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-k49s4"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.992347 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5fdmt"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.992708 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.992787 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2z5sq"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.992962 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.993050 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.995540 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.999377 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rnqdz"] Nov 24 14:21:52 crc kubenswrapper[4822]: I1124 14:21:52.999618 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.000367 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.000476 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.000558 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.000825 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.000870 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.000967 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001004 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001035 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001072 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001106 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001045 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001222 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001242 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001351 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001457 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001548 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001607 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001353 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001393 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001688 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001716 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.001748 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.002583 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.002659 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.003865 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-56tjb"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.024371 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.024537 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.026924 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.027251 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.027403 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.027540 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.031888 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-p9tw4"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.031939 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qr4nb"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.031965 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.032943 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033096 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chmnt\" (UniqueName: \"kubernetes.io/projected/90001d98-f4e4-42f3-8f78-453f244a6286-kube-api-access-chmnt\") pod \"openshift-config-operator-7777fb866f-ctks4\" (UID: \"90001d98-f4e4-42f3-8f78-453f244a6286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033130 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e83d919d-2e04-43b4-9f0c-40aba2d18772-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033165 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-config\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033237 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6jxg\" (UniqueName: \"kubernetes.io/projected/c9acfdd4-def9-4f63-aec3-2f739beae15d-kube-api-access-j6jxg\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033397 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72315396-ab30-4736-8eb9-f735ef0e7f97-serving-cert\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033428 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-client-ca\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033466 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2446a52a-a882-4ebd-972d-eb61d2e6c086-config\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033526 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033552 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/90001d98-f4e4-42f3-8f78-453f244a6286-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ctks4\" (UID: \"90001d98-f4e4-42f3-8f78-453f244a6286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033600 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e83d919d-2e04-43b4-9f0c-40aba2d18772-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033643 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c9acfdd4-def9-4f63-aec3-2f739beae15d-machine-approver-tls\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033685 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-client-ca\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033736 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c9acfdd4-def9-4f63-aec3-2f739beae15d-auth-proxy-config\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033757 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-config\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033776 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033800 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68k2q\" (UniqueName: \"kubernetes.io/projected/397bb803-fafb-4fca-ae9f-e373d413958f-kube-api-access-68k2q\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033819 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033860 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2446a52a-a882-4ebd-972d-eb61d2e6c086-images\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033908 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90001d98-f4e4-42f3-8f78-453f244a6286-serving-cert\") pod \"openshift-config-operator-7777fb866f-ctks4\" (UID: \"90001d98-f4e4-42f3-8f78-453f244a6286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.033936 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034150 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkk4s\" (UniqueName: \"kubernetes.io/projected/72315396-ab30-4736-8eb9-f735ef0e7f97-kube-api-access-hkk4s\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034173 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/397bb803-fafb-4fca-ae9f-e373d413958f-serving-cert\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034194 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034232 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034246 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034300 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034352 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9acfdd4-def9-4f63-aec3-2f739beae15d-config\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034397 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034435 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgd7p\" (UniqueName: \"kubernetes.io/projected/902bd164-754e-4c87-b042-ca87c15d0263-kube-api-access-cgd7p\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034460 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79gf6\" (UniqueName: \"kubernetes.io/projected/2446a52a-a882-4ebd-972d-eb61d2e6c086-kube-api-access-79gf6\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034487 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/902bd164-754e-4c87-b042-ca87c15d0263-audit-dir\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034531 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e83d919d-2e04-43b4-9f0c-40aba2d18772-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034553 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034575 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034595 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2446a52a-a882-4ebd-972d-eb61d2e6c086-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034616 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-audit-policies\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034645 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-674fz\" (UniqueName: \"kubernetes.io/projected/911c455a-0bf1-4855-956f-34a7d3f2f080-kube-api-access-674fz\") pod \"downloads-7954f5f757-p9tw4\" (UID: \"911c455a-0bf1-4855-956f-34a7d3f2f080\") " pod="openshift-console/downloads-7954f5f757-p9tw4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034788 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034808 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gbsg\" (UniqueName: \"kubernetes.io/projected/e83d919d-2e04-43b4-9f0c-40aba2d18772-kube-api-access-6gbsg\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034806 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034919 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.034954 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.036817 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.047072 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.047232 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.047499 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.048074 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.048242 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.049575 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.050179 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.050388 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.051082 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.052473 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.053904 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hl4bh"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.053943 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.054516 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.054657 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.054841 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.055176 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.055540 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.055742 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.057039 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.057854 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-w4gtt"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.058315 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.058697 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.058745 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-55vjd"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.060636 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.060872 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.061047 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.061235 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.061392 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.061683 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.061900 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.062065 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.062429 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.062618 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.063013 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.063473 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.063739 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.063925 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.064651 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.064862 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.064949 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.065036 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.065053 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.065143 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.065171 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.065330 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.068362 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.072819 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.073734 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.075134 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-95knp"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.075557 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.076139 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.085094 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.086950 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.087749 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.101482 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.102104 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.108968 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.112564 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v69sk"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.112881 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.113575 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.120521 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.121896 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.122285 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-k6s8b"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.123185 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.124857 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.125937 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-k49s4"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.128004 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.130597 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.133492 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136501 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72315396-ab30-4736-8eb9-f735ef0e7f97-serving-cert\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136546 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-trusted-ca-bundle\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136573 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15f981aa-3735-40f5-bc13-2cfad43ab9d8-serving-cert\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136592 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/338efc62-c0d6-4589-a6a8-dd783e0f08ac-metrics-certs\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136633 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2446a52a-a882-4ebd-972d-eb61d2e6c086-config\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136662 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c019c6e3-3a0d-4838-a28d-2392d45c52b7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-r2lc6\" (UID: \"c019c6e3-3a0d-4838-a28d-2392d45c52b7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136757 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7d76\" (UniqueName: \"kubernetes.io/projected/195af045-05d5-4240-8115-5772fb13d082-kube-api-access-b7d76\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136777 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/15f981aa-3735-40f5-bc13-2cfad43ab9d8-audit-dir\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136807 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b934905c-f8df-4e00-9926-19df23d81e61-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ksvsx\" (UID: \"b934905c-f8df-4e00-9926-19df23d81e61\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136838 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28e02b8d-b939-4a4c-952e-3e8365a1d124-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5w65w\" (UID: \"28e02b8d-b939-4a4c-952e-3e8365a1d124\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136867 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8d60f83e-1c5e-4839-89c4-fb49928e91a8-metrics-tls\") pod \"dns-operator-744455d44c-rnqdz\" (UID: \"8d60f83e-1c5e-4839-89c4-fb49928e91a8\") " pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.136886 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8593f46e-611a-46a9-9644-aed898afd907-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-m56tg\" (UID: \"8593f46e-611a-46a9-9644-aed898afd907\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.137983 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/70e35241-b7b3-4693-b7bb-a439ee947d03-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-b47sz\" (UID: \"70e35241-b7b3-4693-b7bb-a439ee947d03\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138033 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-serving-cert\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138065 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-oauth-config\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138111 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjfrq\" (UniqueName: \"kubernetes.io/projected/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-kube-api-access-jjfrq\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138139 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-config\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138126 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2446a52a-a882-4ebd-972d-eb61d2e6c086-config\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138175 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zmkp\" (UniqueName: \"kubernetes.io/projected/e49fad7d-a39a-4270-9d25-50d381408838-kube-api-access-9zmkp\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138266 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138342 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e49fad7d-a39a-4270-9d25-50d381408838-audit-dir\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138400 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90001d98-f4e4-42f3-8f78-453f244a6286-serving-cert\") pod \"openshift-config-operator-7777fb866f-ctks4\" (UID: \"90001d98-f4e4-42f3-8f78-453f244a6286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138427 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/195af045-05d5-4240-8115-5772fb13d082-trusted-ca\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138479 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c019c6e3-3a0d-4838-a28d-2392d45c52b7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-r2lc6\" (UID: \"c019c6e3-3a0d-4838-a28d-2392d45c52b7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138509 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-serving-cert\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138558 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138602 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/397bb803-fafb-4fca-ae9f-e373d413958f-serving-cert\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138660 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9acfdd4-def9-4f63-aec3-2f739beae15d-config\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138687 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e49fad7d-a39a-4270-9d25-50d381408838-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138736 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/985f7b3a-f9d1-44cc-9132-6e24bc94e8f3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z2jvf\" (UID: \"985f7b3a-f9d1-44cc-9132-6e24bc94e8f3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138771 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79gf6\" (UniqueName: \"kubernetes.io/projected/2446a52a-a882-4ebd-972d-eb61d2e6c086-kube-api-access-79gf6\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138816 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgd7p\" (UniqueName: \"kubernetes.io/projected/902bd164-754e-4c87-b042-ca87c15d0263-kube-api-access-cgd7p\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138847 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-etcd-service-ca\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138895 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/15f981aa-3735-40f5-bc13-2cfad43ab9d8-node-pullsecrets\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138927 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/902bd164-754e-4c87-b042-ca87c15d0263-audit-dir\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.138989 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/195af045-05d5-4240-8115-5772fb13d082-serving-cert\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139016 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a1a53d5-5871-4b71-8d79-f496da018491-config\") pod \"kube-controller-manager-operator-78b949d7b-m57q2\" (UID: \"8a1a53d5-5871-4b71-8d79-f496da018491\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139063 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139096 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2446a52a-a882-4ebd-972d-eb61d2e6c086-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139162 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-audit-policies\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139190 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpgvq\" (UniqueName: \"kubernetes.io/projected/c019c6e3-3a0d-4838-a28d-2392d45c52b7-kube-api-access-cpgvq\") pod \"kube-storage-version-migrator-operator-b67b599dd-r2lc6\" (UID: \"c019c6e3-3a0d-4838-a28d-2392d45c52b7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139255 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139280 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a1a53d5-5871-4b71-8d79-f496da018491-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-m57q2\" (UID: \"8a1a53d5-5871-4b71-8d79-f496da018491\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139328 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d5de9e-3cc2-46a7-8138-cca5157799e5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xtzhk\" (UID: \"02d5de9e-3cc2-46a7-8138-cca5157799e5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139358 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-service-ca\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139432 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e49fad7d-a39a-4270-9d25-50d381408838-serving-cert\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139460 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/70e35241-b7b3-4693-b7bb-a439ee947d03-proxy-tls\") pod \"machine-config-controller-84d6567774-b47sz\" (UID: \"70e35241-b7b3-4693-b7bb-a439ee947d03\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139510 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139541 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-metrics-tls\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139583 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-etcd-client\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139608 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qfl9\" (UniqueName: \"kubernetes.io/projected/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-kube-api-access-4qfl9\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139628 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e49fad7d-a39a-4270-9d25-50d381408838-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139678 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-config\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139696 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d5de9e-3cc2-46a7-8138-cca5157799e5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xtzhk\" (UID: \"02d5de9e-3cc2-46a7-8138-cca5157799e5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139744 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d58f06c7-b1a1-48ae-8b56-78b7f08970f5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-g7jjl\" (UID: \"d58f06c7-b1a1-48ae-8b56-78b7f08970f5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139769 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw8tx\" (UniqueName: \"kubernetes.io/projected/02d5de9e-3cc2-46a7-8138-cca5157799e5-kube-api-access-hw8tx\") pod \"openshift-apiserver-operator-796bbdcf4f-xtzhk\" (UID: \"02d5de9e-3cc2-46a7-8138-cca5157799e5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.139791 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/15f981aa-3735-40f5-bc13-2cfad43ab9d8-etcd-client\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140244 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-image-import-ca\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140305 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klb2l\" (UniqueName: \"kubernetes.io/projected/338efc62-c0d6-4589-a6a8-dd783e0f08ac-kube-api-access-klb2l\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140386 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-client-ca\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140414 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dvgh\" (UniqueName: \"kubernetes.io/projected/6153af6d-c527-485f-8a0e-0cfecfb43d5b-kube-api-access-6dvgh\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140705 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140736 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/90001d98-f4e4-42f3-8f78-453f244a6286-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ctks4\" (UID: \"90001d98-f4e4-42f3-8f78-453f244a6286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140760 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/195af045-05d5-4240-8115-5772fb13d082-config\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140811 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e83d919d-2e04-43b4-9f0c-40aba2d18772-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140835 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d58f06c7-b1a1-48ae-8b56-78b7f08970f5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-g7jjl\" (UID: \"d58f06c7-b1a1-48ae-8b56-78b7f08970f5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140878 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgzb5\" (UniqueName: \"kubernetes.io/projected/985f7b3a-f9d1-44cc-9132-6e24bc94e8f3-kube-api-access-mgzb5\") pod \"cluster-samples-operator-665b6dd947-z2jvf\" (UID: \"985f7b3a-f9d1-44cc-9132-6e24bc94e8f3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140905 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lznbh\" (UniqueName: \"kubernetes.io/projected/6b8514e2-4241-4916-b843-d62897d4af2c-kube-api-access-lznbh\") pod \"control-plane-machine-set-operator-78cbb6b69f-vqxq9\" (UID: \"6b8514e2-4241-4916-b843-d62897d4af2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140953 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-config\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.140979 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-serving-cert\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141024 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c9acfdd4-def9-4f63-aec3-2f739beae15d-machine-approver-tls\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141047 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/338efc62-c0d6-4589-a6a8-dd783e0f08ac-service-ca-bundle\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141070 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-service-ca-bundle\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141120 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-client-ca\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141145 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c9acfdd4-def9-4f63-aec3-2f739beae15d-auth-proxy-config\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141433 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mqrf\" (UniqueName: \"kubernetes.io/projected/b934905c-f8df-4e00-9926-19df23d81e61-kube-api-access-7mqrf\") pod \"package-server-manager-789f6589d5-ksvsx\" (UID: \"b934905c-f8df-4e00-9926-19df23d81e61\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141467 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141492 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68k2q\" (UniqueName: \"kubernetes.io/projected/397bb803-fafb-4fca-ae9f-e373d413958f-kube-api-access-68k2q\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141528 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/15f981aa-3735-40f5-bc13-2cfad43ab9d8-encryption-config\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141550 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvh7d\" (UniqueName: \"kubernetes.io/projected/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-kube-api-access-rvh7d\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141579 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2446a52a-a882-4ebd-972d-eb61d2e6c086-images\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141603 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-oauth-serving-cert\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141669 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-trusted-ca\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141700 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkk4s\" (UniqueName: \"kubernetes.io/projected/72315396-ab30-4736-8eb9-f735ef0e7f97-kube-api-access-hkk4s\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141722 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b8514e2-4241-4916-b843-d62897d4af2c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vqxq9\" (UID: \"6b8514e2-4241-4916-b843-d62897d4af2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141748 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28e02b8d-b939-4a4c-952e-3e8365a1d124-config\") pod \"kube-apiserver-operator-766d6c64bb-5w65w\" (UID: \"28e02b8d-b939-4a4c-952e-3e8365a1d124\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141774 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8593f46e-611a-46a9-9644-aed898afd907-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-m56tg\" (UID: \"8593f46e-611a-46a9-9644-aed898afd907\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141797 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx9qh\" (UniqueName: \"kubernetes.io/projected/15f981aa-3735-40f5-bc13-2cfad43ab9d8-kube-api-access-jx9qh\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141828 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141847 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141870 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141892 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-config\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141914 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-etcd-serving-ca\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141963 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.141991 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e49fad7d-a39a-4270-9d25-50d381408838-audit-policies\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.142036 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.142068 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/338efc62-c0d6-4589-a6a8-dd783e0f08ac-default-certificate\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.142116 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e83d919d-2e04-43b4-9f0c-40aba2d18772-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.142139 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e49fad7d-a39a-4270-9d25-50d381408838-etcd-client\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.142153 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-config\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.142164 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.142234 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e49fad7d-a39a-4270-9d25-50d381408838-encryption-config\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.142263 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-config\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.144476 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e83d919d-2e04-43b4-9f0c-40aba2d18772-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.145331 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2446a52a-a882-4ebd-972d-eb61d2e6c086-images\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.146294 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c9acfdd4-def9-4f63-aec3-2f739beae15d-auth-proxy-config\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.146405 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9acfdd4-def9-4f63-aec3-2f739beae15d-config\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.146753 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.146787 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-audit-policies\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.146739 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/90001d98-f4e4-42f3-8f78-453f244a6286-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ctks4\" (UID: \"90001d98-f4e4-42f3-8f78-453f244a6286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.146824 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.148181 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/902bd164-754e-4c87-b042-ca87c15d0263-audit-dir\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.148472 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.148899 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-client-ca\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.148982 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72315396-ab30-4736-8eb9-f735ef0e7f97-serving-cert\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.150631 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.150763 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.150869 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.151964 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-config\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.152683 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c9acfdd4-def9-4f63-aec3-2f739beae15d-machine-approver-tls\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153024 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153161 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153428 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153535 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e83d919d-2e04-43b4-9f0c-40aba2d18772-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.144985 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90001d98-f4e4-42f3-8f78-453f244a6286-serving-cert\") pod \"openshift-config-operator-7777fb866f-ctks4\" (UID: \"90001d98-f4e4-42f3-8f78-453f244a6286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153587 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-client-ca\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153643 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-audit\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153677 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8a1a53d5-5871-4b71-8d79-f496da018491-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-m57q2\" (UID: \"8a1a53d5-5871-4b71-8d79-f496da018491\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153706 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppzqn\" (UniqueName: \"kubernetes.io/projected/8593f46e-611a-46a9-9644-aed898afd907-kube-api-access-ppzqn\") pod \"openshift-controller-manager-operator-756b6f6bc6-m56tg\" (UID: \"8593f46e-611a-46a9-9644-aed898afd907\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153734 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gbsg\" (UniqueName: \"kubernetes.io/projected/e83d919d-2e04-43b4-9f0c-40aba2d18772-kube-api-access-6gbsg\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153744 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153753 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-674fz\" (UniqueName: \"kubernetes.io/projected/911c455a-0bf1-4855-956f-34a7d3f2f080-kube-api-access-674fz\") pod \"downloads-7954f5f757-p9tw4\" (UID: \"911c455a-0bf1-4855-956f-34a7d3f2f080\") " pod="openshift-console/downloads-7954f5f757-p9tw4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153771 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/338efc62-c0d6-4589-a6a8-dd783e0f08ac-stats-auth\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153795 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28e02b8d-b939-4a4c-952e-3e8365a1d124-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5w65w\" (UID: \"28e02b8d-b939-4a4c-952e-3e8365a1d124\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153814 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2bf05124-0e4f-4262-95c6-d87ede36dcb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-55vjd\" (UID: \"2bf05124-0e4f-4262-95c6-d87ede36dcb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153831 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153854 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58f06c7-b1a1-48ae-8b56-78b7f08970f5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-g7jjl\" (UID: \"d58f06c7-b1a1-48ae-8b56-78b7f08970f5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153880 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chmnt\" (UniqueName: \"kubernetes.io/projected/90001d98-f4e4-42f3-8f78-453f244a6286-kube-api-access-chmnt\") pod \"openshift-config-operator-7777fb866f-ctks4\" (UID: \"90001d98-f4e4-42f3-8f78-453f244a6286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153901 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e83d919d-2e04-43b4-9f0c-40aba2d18772-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153927 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-config\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153948 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-etcd-ca\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153969 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgckk\" (UniqueName: \"kubernetes.io/projected/70e35241-b7b3-4693-b7bb-a439ee947d03-kube-api-access-zgckk\") pod \"machine-config-controller-84d6567774-b47sz\" (UID: \"70e35241-b7b3-4693-b7bb-a439ee947d03\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.154027 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq47m\" (UniqueName: \"kubernetes.io/projected/2bf05124-0e4f-4262-95c6-d87ede36dcb5-kube-api-access-mq47m\") pod \"multus-admission-controller-857f4d67dd-55vjd\" (UID: \"2bf05124-0e4f-4262-95c6-d87ede36dcb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.154054 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6jxg\" (UniqueName: \"kubernetes.io/projected/c9acfdd4-def9-4f63-aec3-2f739beae15d-kube-api-access-j6jxg\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.154079 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2j2s\" (UniqueName: \"kubernetes.io/projected/0d499682-90a9-464e-9cb5-6c5a82155311-kube-api-access-b2j2s\") pod \"migrator-59844c95c7-rmhjj\" (UID: \"0d499682-90a9-464e-9cb5-6c5a82155311\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.154102 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk7dd\" (UniqueName: \"kubernetes.io/projected/8d60f83e-1c5e-4839-89c4-fb49928e91a8-kube-api-access-lk7dd\") pod \"dns-operator-744455d44c-rnqdz\" (UID: \"8d60f83e-1c5e-4839-89c4-fb49928e91a8\") " pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.154410 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.153900 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.154537 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5v7wp"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.154637 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/397bb803-fafb-4fca-ae9f-e373d413958f-serving-cert\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.154666 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.154928 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.155768 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2446a52a-a882-4ebd-972d-eb61d2e6c086-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.156326 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.157758 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rnqdz"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.159111 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8bc7n"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.160500 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gzxq"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.161922 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.164631 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.164972 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.165143 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-55vjd"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.166432 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.167290 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ctks4"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.168366 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-56tjb"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.171817 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.173049 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2z5sq"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.174140 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.175460 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5fdmt"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.176705 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dq7pg"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.178608 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.178805 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.180093 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6rhjd"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.182701 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.182838 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.182892 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.184528 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.184593 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.185420 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.186641 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.187708 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.188966 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.190098 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.192101 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.192122 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.193150 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dq7pg"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.195238 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-95knp"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.196584 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.197848 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v69sk"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.198868 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6rhjd"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.199840 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qldnw"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.200794 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qldnw" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.202252 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qldnw"] Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.204942 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.225310 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.244746 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.254914 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a1a53d5-5871-4b71-8d79-f496da018491-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-m57q2\" (UID: \"8a1a53d5-5871-4b71-8d79-f496da018491\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.254947 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d5de9e-3cc2-46a7-8138-cca5157799e5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xtzhk\" (UID: \"02d5de9e-3cc2-46a7-8138-cca5157799e5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.254967 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-service-ca\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.254983 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e49fad7d-a39a-4270-9d25-50d381408838-serving-cert\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.254999 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/70e35241-b7b3-4693-b7bb-a439ee947d03-proxy-tls\") pod \"machine-config-controller-84d6567774-b47sz\" (UID: \"70e35241-b7b3-4693-b7bb-a439ee947d03\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255015 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-metrics-tls\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255033 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-etcd-client\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255055 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qfl9\" (UniqueName: \"kubernetes.io/projected/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-kube-api-access-4qfl9\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255070 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e49fad7d-a39a-4270-9d25-50d381408838-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255085 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d5de9e-3cc2-46a7-8138-cca5157799e5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xtzhk\" (UID: \"02d5de9e-3cc2-46a7-8138-cca5157799e5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255108 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d58f06c7-b1a1-48ae-8b56-78b7f08970f5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-g7jjl\" (UID: \"d58f06c7-b1a1-48ae-8b56-78b7f08970f5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255123 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw8tx\" (UniqueName: \"kubernetes.io/projected/02d5de9e-3cc2-46a7-8138-cca5157799e5-kube-api-access-hw8tx\") pod \"openshift-apiserver-operator-796bbdcf4f-xtzhk\" (UID: \"02d5de9e-3cc2-46a7-8138-cca5157799e5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255139 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/15f981aa-3735-40f5-bc13-2cfad43ab9d8-etcd-client\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255153 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-image-import-ca\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255173 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klb2l\" (UniqueName: \"kubernetes.io/projected/338efc62-c0d6-4589-a6a8-dd783e0f08ac-kube-api-access-klb2l\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255188 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dvgh\" (UniqueName: \"kubernetes.io/projected/6153af6d-c527-485f-8a0e-0cfecfb43d5b-kube-api-access-6dvgh\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255241 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/195af045-05d5-4240-8115-5772fb13d082-config\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255265 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d58f06c7-b1a1-48ae-8b56-78b7f08970f5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-g7jjl\" (UID: \"d58f06c7-b1a1-48ae-8b56-78b7f08970f5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255311 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgzb5\" (UniqueName: \"kubernetes.io/projected/985f7b3a-f9d1-44cc-9132-6e24bc94e8f3-kube-api-access-mgzb5\") pod \"cluster-samples-operator-665b6dd947-z2jvf\" (UID: \"985f7b3a-f9d1-44cc-9132-6e24bc94e8f3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255333 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lznbh\" (UniqueName: \"kubernetes.io/projected/6b8514e2-4241-4916-b843-d62897d4af2c-kube-api-access-lznbh\") pod \"control-plane-machine-set-operator-78cbb6b69f-vqxq9\" (UID: \"6b8514e2-4241-4916-b843-d62897d4af2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255349 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-config\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255367 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-serving-cert\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255382 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/338efc62-c0d6-4589-a6a8-dd783e0f08ac-service-ca-bundle\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255400 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-service-ca-bundle\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255418 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mqrf\" (UniqueName: \"kubernetes.io/projected/b934905c-f8df-4e00-9926-19df23d81e61-kube-api-access-7mqrf\") pod \"package-server-manager-789f6589d5-ksvsx\" (UID: \"b934905c-f8df-4e00-9926-19df23d81e61\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255452 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/15f981aa-3735-40f5-bc13-2cfad43ab9d8-encryption-config\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255470 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvh7d\" (UniqueName: \"kubernetes.io/projected/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-kube-api-access-rvh7d\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255487 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-oauth-serving-cert\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255502 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-trusted-ca\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255524 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b8514e2-4241-4916-b843-d62897d4af2c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vqxq9\" (UID: \"6b8514e2-4241-4916-b843-d62897d4af2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255543 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28e02b8d-b939-4a4c-952e-3e8365a1d124-config\") pod \"kube-apiserver-operator-766d6c64bb-5w65w\" (UID: \"28e02b8d-b939-4a4c-952e-3e8365a1d124\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255559 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8593f46e-611a-46a9-9644-aed898afd907-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-m56tg\" (UID: \"8593f46e-611a-46a9-9644-aed898afd907\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255575 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx9qh\" (UniqueName: \"kubernetes.io/projected/15f981aa-3735-40f5-bc13-2cfad43ab9d8-kube-api-access-jx9qh\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255592 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-config\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255608 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-etcd-serving-ca\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255621 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e49fad7d-a39a-4270-9d25-50d381408838-audit-policies\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255637 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255652 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/338efc62-c0d6-4589-a6a8-dd783e0f08ac-default-certificate\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255668 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e49fad7d-a39a-4270-9d25-50d381408838-etcd-client\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255684 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e49fad7d-a39a-4270-9d25-50d381408838-encryption-config\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255698 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-config\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255714 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-audit\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255730 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8a1a53d5-5871-4b71-8d79-f496da018491-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-m57q2\" (UID: \"8a1a53d5-5871-4b71-8d79-f496da018491\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255747 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppzqn\" (UniqueName: \"kubernetes.io/projected/8593f46e-611a-46a9-9644-aed898afd907-kube-api-access-ppzqn\") pod \"openshift-controller-manager-operator-756b6f6bc6-m56tg\" (UID: \"8593f46e-611a-46a9-9644-aed898afd907\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255774 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/338efc62-c0d6-4589-a6a8-dd783e0f08ac-stats-auth\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255792 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28e02b8d-b939-4a4c-952e-3e8365a1d124-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5w65w\" (UID: \"28e02b8d-b939-4a4c-952e-3e8365a1d124\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255813 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2bf05124-0e4f-4262-95c6-d87ede36dcb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-55vjd\" (UID: \"2bf05124-0e4f-4262-95c6-d87ede36dcb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255835 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.255910 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58f06c7-b1a1-48ae-8b56-78b7f08970f5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-g7jjl\" (UID: \"d58f06c7-b1a1-48ae-8b56-78b7f08970f5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.257765 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-image-import-ca\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.257755 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.258249 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/195af045-05d5-4240-8115-5772fb13d082-config\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.258545 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e49fad7d-a39a-4270-9d25-50d381408838-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.259031 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-oauth-serving-cert\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.259110 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-config\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.259970 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d5de9e-3cc2-46a7-8138-cca5157799e5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xtzhk\" (UID: \"02d5de9e-3cc2-46a7-8138-cca5157799e5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.260601 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-service-ca\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.268396 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/15f981aa-3735-40f5-bc13-2cfad43ab9d8-encryption-config\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.268824 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.268866 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-etcd-client\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.268983 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8593f46e-611a-46a9-9644-aed898afd907-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-m56tg\" (UID: \"8593f46e-611a-46a9-9644-aed898afd907\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.270030 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-service-ca-bundle\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.275098 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-config\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.277699 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e49fad7d-a39a-4270-9d25-50d381408838-encryption-config\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.278117 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e49fad7d-a39a-4270-9d25-50d381408838-audit-policies\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.278458 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/15f981aa-3735-40f5-bc13-2cfad43ab9d8-etcd-client\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.278502 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-config\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.278797 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-serving-cert\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.279086 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e49fad7d-a39a-4270-9d25-50d381408838-serving-cert\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.280356 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d5de9e-3cc2-46a7-8138-cca5157799e5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xtzhk\" (UID: \"02d5de9e-3cc2-46a7-8138-cca5157799e5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.282996 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-audit\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.283546 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/15f981aa-3735-40f5-bc13-2cfad43ab9d8-etcd-serving-ca\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.285474 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-config\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.285556 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-etcd-ca\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.285686 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgckk\" (UniqueName: \"kubernetes.io/projected/70e35241-b7b3-4693-b7bb-a439ee947d03-kube-api-access-zgckk\") pod \"machine-config-controller-84d6567774-b47sz\" (UID: \"70e35241-b7b3-4693-b7bb-a439ee947d03\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.285720 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq47m\" (UniqueName: \"kubernetes.io/projected/2bf05124-0e4f-4262-95c6-d87ede36dcb5-kube-api-access-mq47m\") pod \"multus-admission-controller-857f4d67dd-55vjd\" (UID: \"2bf05124-0e4f-4262-95c6-d87ede36dcb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.285940 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.286304 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2j2s\" (UniqueName: \"kubernetes.io/projected/0d499682-90a9-464e-9cb5-6c5a82155311-kube-api-access-b2j2s\") pod \"migrator-59844c95c7-rmhjj\" (UID: \"0d499682-90a9-464e-9cb5-6c5a82155311\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.286445 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-etcd-ca\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.286462 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk7dd\" (UniqueName: \"kubernetes.io/projected/8d60f83e-1c5e-4839-89c4-fb49928e91a8-kube-api-access-lk7dd\") pod \"dns-operator-744455d44c-rnqdz\" (UID: \"8d60f83e-1c5e-4839-89c4-fb49928e91a8\") " pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.287797 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-trusted-ca-bundle\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288153 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15f981aa-3735-40f5-bc13-2cfad43ab9d8-serving-cert\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288199 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/338efc62-c0d6-4589-a6a8-dd783e0f08ac-metrics-certs\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288298 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c019c6e3-3a0d-4838-a28d-2392d45c52b7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-r2lc6\" (UID: \"c019c6e3-3a0d-4838-a28d-2392d45c52b7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288327 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7d76\" (UniqueName: \"kubernetes.io/projected/195af045-05d5-4240-8115-5772fb13d082-kube-api-access-b7d76\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288346 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/15f981aa-3735-40f5-bc13-2cfad43ab9d8-audit-dir\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288363 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b934905c-f8df-4e00-9926-19df23d81e61-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ksvsx\" (UID: \"b934905c-f8df-4e00-9926-19df23d81e61\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288381 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28e02b8d-b939-4a4c-952e-3e8365a1d124-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5w65w\" (UID: \"28e02b8d-b939-4a4c-952e-3e8365a1d124\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288400 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8d60f83e-1c5e-4839-89c4-fb49928e91a8-metrics-tls\") pod \"dns-operator-744455d44c-rnqdz\" (UID: \"8d60f83e-1c5e-4839-89c4-fb49928e91a8\") " pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288418 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8593f46e-611a-46a9-9644-aed898afd907-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-m56tg\" (UID: \"8593f46e-611a-46a9-9644-aed898afd907\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288436 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/70e35241-b7b3-4693-b7bb-a439ee947d03-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-b47sz\" (UID: \"70e35241-b7b3-4693-b7bb-a439ee947d03\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288433 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/15f981aa-3735-40f5-bc13-2cfad43ab9d8-audit-dir\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288453 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-serving-cert\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288470 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-oauth-config\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288499 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjfrq\" (UniqueName: \"kubernetes.io/projected/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-kube-api-access-jjfrq\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288518 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zmkp\" (UniqueName: \"kubernetes.io/projected/e49fad7d-a39a-4270-9d25-50d381408838-kube-api-access-9zmkp\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288539 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e49fad7d-a39a-4270-9d25-50d381408838-audit-dir\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288580 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/195af045-05d5-4240-8115-5772fb13d082-trusted-ca\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288596 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c019c6e3-3a0d-4838-a28d-2392d45c52b7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-r2lc6\" (UID: \"c019c6e3-3a0d-4838-a28d-2392d45c52b7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288610 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-serving-cert\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288633 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e49fad7d-a39a-4270-9d25-50d381408838-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288651 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/985f7b3a-f9d1-44cc-9132-6e24bc94e8f3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z2jvf\" (UID: \"985f7b3a-f9d1-44cc-9132-6e24bc94e8f3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288678 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-etcd-service-ca\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288696 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/15f981aa-3735-40f5-bc13-2cfad43ab9d8-node-pullsecrets\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288713 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/195af045-05d5-4240-8115-5772fb13d082-serving-cert\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288730 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a1a53d5-5871-4b71-8d79-f496da018491-config\") pod \"kube-controller-manager-operator-78b949d7b-m57q2\" (UID: \"8a1a53d5-5871-4b71-8d79-f496da018491\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288746 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288769 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpgvq\" (UniqueName: \"kubernetes.io/projected/c019c6e3-3a0d-4838-a28d-2392d45c52b7-kube-api-access-cpgvq\") pod \"kube-storage-version-migrator-operator-b67b599dd-r2lc6\" (UID: \"c019c6e3-3a0d-4838-a28d-2392d45c52b7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.288906 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e49fad7d-a39a-4270-9d25-50d381408838-audit-dir\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.289603 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-trusted-ca-bundle\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.289728 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/195af045-05d5-4240-8115-5772fb13d082-trusted-ca\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.289786 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/15f981aa-3735-40f5-bc13-2cfad43ab9d8-node-pullsecrets\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.290014 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-etcd-service-ca\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.290151 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/70e35241-b7b3-4693-b7bb-a439ee947d03-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-b47sz\" (UID: \"70e35241-b7b3-4693-b7bb-a439ee947d03\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.290937 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8593f46e-611a-46a9-9644-aed898afd907-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-m56tg\" (UID: \"8593f46e-611a-46a9-9644-aed898afd907\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.290998 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.291627 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15f981aa-3735-40f5-bc13-2cfad43ab9d8-serving-cert\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.292004 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/195af045-05d5-4240-8115-5772fb13d082-serving-cert\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.292242 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e49fad7d-a39a-4270-9d25-50d381408838-etcd-client\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.292428 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e49fad7d-a39a-4270-9d25-50d381408838-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.293252 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-oauth-config\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.293668 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-serving-cert\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.293995 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/985f7b3a-f9d1-44cc-9132-6e24bc94e8f3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z2jvf\" (UID: \"985f7b3a-f9d1-44cc-9132-6e24bc94e8f3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.295957 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-serving-cert\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.296112 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8d60f83e-1c5e-4839-89c4-fb49928e91a8-metrics-tls\") pod \"dns-operator-744455d44c-rnqdz\" (UID: \"8d60f83e-1c5e-4839-89c4-fb49928e91a8\") " pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.304723 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.327343 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.327908 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-config\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.344686 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.368853 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.385050 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.405362 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.413241 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d58f06c7-b1a1-48ae-8b56-78b7f08970f5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-g7jjl\" (UID: \"d58f06c7-b1a1-48ae-8b56-78b7f08970f5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.424321 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.426053 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58f06c7-b1a1-48ae-8b56-78b7f08970f5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-g7jjl\" (UID: \"d58f06c7-b1a1-48ae-8b56-78b7f08970f5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.451226 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.461090 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-trusted-ca\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.464959 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.484278 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.491803 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-metrics-tls\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.505387 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.525625 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.545290 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.553594 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/70e35241-b7b3-4693-b7bb-a439ee947d03-proxy-tls\") pod \"machine-config-controller-84d6567774-b47sz\" (UID: \"70e35241-b7b3-4693-b7bb-a439ee947d03\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.565446 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.584381 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.605316 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.625794 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.646282 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.652444 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c019c6e3-3a0d-4838-a28d-2392d45c52b7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-r2lc6\" (UID: \"c019c6e3-3a0d-4838-a28d-2392d45c52b7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.665370 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.685687 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.692492 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c019c6e3-3a0d-4838-a28d-2392d45c52b7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-r2lc6\" (UID: \"c019c6e3-3a0d-4838-a28d-2392d45c52b7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.705460 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.715772 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28e02b8d-b939-4a4c-952e-3e8365a1d124-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5w65w\" (UID: \"28e02b8d-b939-4a4c-952e-3e8365a1d124\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.724839 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.745069 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.750820 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28e02b8d-b939-4a4c-952e-3e8365a1d124-config\") pod \"kube-apiserver-operator-766d6c64bb-5w65w\" (UID: \"28e02b8d-b939-4a4c-952e-3e8365a1d124\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.764423 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.784868 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.792775 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/338efc62-c0d6-4589-a6a8-dd783e0f08ac-stats-auth\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.805132 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.815148 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/338efc62-c0d6-4589-a6a8-dd783e0f08ac-metrics-certs\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.827515 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.843978 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.851072 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/338efc62-c0d6-4589-a6a8-dd783e0f08ac-service-ca-bundle\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.865122 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.884881 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.904957 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.914702 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/338efc62-c0d6-4589-a6a8-dd783e0f08ac-default-certificate\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.924317 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.945274 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.965812 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.984622 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 14:21:53 crc kubenswrapper[4822]: I1124 14:21:53.995165 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2bf05124-0e4f-4262-95c6-d87ede36dcb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-55vjd\" (UID: \"2bf05124-0e4f-4262-95c6-d87ede36dcb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.005550 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.026357 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.045379 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.063054 4822 request.go:700] Waited for 1.000454996s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.064910 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.085408 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.102043 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a1a53d5-5871-4b71-8d79-f496da018491-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-m57q2\" (UID: \"8a1a53d5-5871-4b71-8d79-f496da018491\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.105446 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.112106 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a1a53d5-5871-4b71-8d79-f496da018491-config\") pod \"kube-controller-manager-operator-78b949d7b-m57q2\" (UID: \"8a1a53d5-5871-4b71-8d79-f496da018491\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.125078 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.145795 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.164818 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.174370 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b8514e2-4241-4916-b843-d62897d4af2c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vqxq9\" (UID: \"6b8514e2-4241-4916-b843-d62897d4af2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.185726 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.204741 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.225160 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.233600 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b934905c-f8df-4e00-9926-19df23d81e61-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ksvsx\" (UID: \"b934905c-f8df-4e00-9926-19df23d81e61\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.245675 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.285287 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.306016 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.325527 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.345388 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.365417 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.385864 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.417725 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.424760 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.445972 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.465908 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.485446 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.506904 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.525252 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.545796 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.565412 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.585799 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.625853 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.645394 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.666252 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.686296 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.704938 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.724419 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.745858 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.765111 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.813725 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68k2q\" (UniqueName: \"kubernetes.io/projected/397bb803-fafb-4fca-ae9f-e373d413958f-kube-api-access-68k2q\") pod \"route-controller-manager-6576b87f9c-q7wf8\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.825698 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkk4s\" (UniqueName: \"kubernetes.io/projected/72315396-ab30-4736-8eb9-f735ef0e7f97-kube-api-access-hkk4s\") pod \"controller-manager-879f6c89f-hl4bh\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.847098 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79gf6\" (UniqueName: \"kubernetes.io/projected/2446a52a-a882-4ebd-972d-eb61d2e6c086-kube-api-access-79gf6\") pod \"machine-api-operator-5694c8668f-qr4nb\" (UID: \"2446a52a-a882-4ebd-972d-eb61d2e6c086\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.863002 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgd7p\" (UniqueName: \"kubernetes.io/projected/902bd164-754e-4c87-b042-ca87c15d0263-kube-api-access-cgd7p\") pod \"oauth-openshift-558db77b4-7gzxq\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.874258 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.897059 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gbsg\" (UniqueName: \"kubernetes.io/projected/e83d919d-2e04-43b4-9f0c-40aba2d18772-kube-api-access-6gbsg\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.909143 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-674fz\" (UniqueName: \"kubernetes.io/projected/911c455a-0bf1-4855-956f-34a7d3f2f080-kube-api-access-674fz\") pod \"downloads-7954f5f757-p9tw4\" (UID: \"911c455a-0bf1-4855-956f-34a7d3f2f080\") " pod="openshift-console/downloads-7954f5f757-p9tw4" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.933618 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chmnt\" (UniqueName: \"kubernetes.io/projected/90001d98-f4e4-42f3-8f78-453f244a6286-kube-api-access-chmnt\") pod \"openshift-config-operator-7777fb866f-ctks4\" (UID: \"90001d98-f4e4-42f3-8f78-453f244a6286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.952533 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6jxg\" (UniqueName: \"kubernetes.io/projected/c9acfdd4-def9-4f63-aec3-2f739beae15d-kube-api-access-j6jxg\") pod \"machine-approver-56656f9798-mscgb\" (UID: \"c9acfdd4-def9-4f63-aec3-2f739beae15d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.965637 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.968519 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e83d919d-2e04-43b4-9f0c-40aba2d18772-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-thks6\" (UID: \"e83d919d-2e04-43b4-9f0c-40aba2d18772\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:54 crc kubenswrapper[4822]: I1124 14:21:54.985913 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.006060 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.018824 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.025125 4822 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.038772 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hl4bh"] Nov 24 14:21:55 crc kubenswrapper[4822]: W1124 14:21:55.044710 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72315396_ab30_4736_8eb9_f735ef0e7f97.slice/crio-3a9098f789c9bc1afb342643ff741d594840f520c39b017051d9b747fd8dbcdb WatchSource:0}: Error finding container 3a9098f789c9bc1afb342643ff741d594840f520c39b017051d9b747fd8dbcdb: Status 404 returned error can't find the container with id 3a9098f789c9bc1afb342643ff741d594840f520c39b017051d9b747fd8dbcdb Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.046147 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.063766 4822 request.go:700] Waited for 1.880586128s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.065767 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.073476 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-p9tw4" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.084716 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.086966 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.096472 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.106535 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.126319 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.135525 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.145254 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.153857 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.198643 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.199454 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qfl9\" (UniqueName: \"kubernetes.io/projected/58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b-kube-api-access-4qfl9\") pod \"etcd-operator-b45778765-56tjb\" (UID: \"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.204755 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klb2l\" (UniqueName: \"kubernetes.io/projected/338efc62-c0d6-4589-a6a8-dd783e0f08ac-kube-api-access-klb2l\") pod \"router-default-5444994796-w4gtt\" (UID: \"338efc62-c0d6-4589-a6a8-dd783e0f08ac\") " pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.220809 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dvgh\" (UniqueName: \"kubernetes.io/projected/6153af6d-c527-485f-8a0e-0cfecfb43d5b-kube-api-access-6dvgh\") pod \"console-f9d7485db-5v7wp\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.235060 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.245466 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvh7d\" (UniqueName: \"kubernetes.io/projected/3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12-kube-api-access-rvh7d\") pod \"authentication-operator-69f744f599-5fdmt\" (UID: \"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.264896 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx9qh\" (UniqueName: \"kubernetes.io/projected/15f981aa-3735-40f5-bc13-2cfad43ab9d8-kube-api-access-jx9qh\") pod \"apiserver-76f77b778f-8bc7n\" (UID: \"15f981aa-3735-40f5-bc13-2cfad43ab9d8\") " pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.277487 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.283339 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d58f06c7-b1a1-48ae-8b56-78b7f08970f5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-g7jjl\" (UID: \"d58f06c7-b1a1-48ae-8b56-78b7f08970f5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.284374 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qr4nb"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.301583 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgzb5\" (UniqueName: \"kubernetes.io/projected/985f7b3a-f9d1-44cc-9132-6e24bc94e8f3-kube-api-access-mgzb5\") pod \"cluster-samples-operator-665b6dd947-z2jvf\" (UID: \"985f7b3a-f9d1-44cc-9132-6e24bc94e8f3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" Nov 24 14:21:55 crc kubenswrapper[4822]: W1124 14:21:55.305264 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2446a52a_a882_4ebd_972d_eb61d2e6c086.slice/crio-c4de0e734d9b5f5e27616d4b5fef21ed4d797dac1b5068bc3b7373949f0eba35 WatchSource:0}: Error finding container c4de0e734d9b5f5e27616d4b5fef21ed4d797dac1b5068bc3b7373949f0eba35: Status 404 returned error can't find the container with id c4de0e734d9b5f5e27616d4b5fef21ed4d797dac1b5068bc3b7373949f0eba35 Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.306943 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.312321 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.323935 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lznbh\" (UniqueName: \"kubernetes.io/projected/6b8514e2-4241-4916-b843-d62897d4af2c-kube-api-access-lznbh\") pod \"control-plane-machine-set-operator-78cbb6b69f-vqxq9\" (UID: \"6b8514e2-4241-4916-b843-d62897d4af2c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.343099 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw8tx\" (UniqueName: \"kubernetes.io/projected/02d5de9e-3cc2-46a7-8138-cca5157799e5-kube-api-access-hw8tx\") pod \"openshift-apiserver-operator-796bbdcf4f-xtzhk\" (UID: \"02d5de9e-3cc2-46a7-8138-cca5157799e5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.343403 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.366276 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8a1a53d5-5871-4b71-8d79-f496da018491-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-m57q2\" (UID: \"8a1a53d5-5871-4b71-8d79-f496da018491\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.383302 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28e02b8d-b939-4a4c-952e-3e8365a1d124-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5w65w\" (UID: \"28e02b8d-b939-4a4c-952e-3e8365a1d124\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.405282 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-p9tw4"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.406333 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.409506 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppzqn\" (UniqueName: \"kubernetes.io/projected/8593f46e-611a-46a9-9644-aed898afd907-kube-api-access-ppzqn\") pod \"openshift-controller-manager-operator-756b6f6bc6-m56tg\" (UID: \"8593f46e-611a-46a9-9644-aed898afd907\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.411841 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.421485 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mqrf\" (UniqueName: \"kubernetes.io/projected/b934905c-f8df-4e00-9926-19df23d81e61-kube-api-access-7mqrf\") pod \"package-server-manager-789f6589d5-ksvsx\" (UID: \"b934905c-f8df-4e00-9926-19df23d81e61\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.439510 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:55 crc kubenswrapper[4822]: W1124 14:21:55.448792 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod911c455a_0bf1_4855_956f_34a7d3f2f080.slice/crio-627c82d09bd6c9722e93ee80d668ac3b67eab5e3aaf0b4f6f48b5df97e990c5f WatchSource:0}: Error finding container 627c82d09bd6c9722e93ee80d668ac3b67eab5e3aaf0b4f6f48b5df97e990c5f: Status 404 returned error can't find the container with id 627c82d09bd6c9722e93ee80d668ac3b67eab5e3aaf0b4f6f48b5df97e990c5f Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.461280 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgckk\" (UniqueName: \"kubernetes.io/projected/70e35241-b7b3-4693-b7bb-a439ee947d03-kube-api-access-zgckk\") pod \"machine-config-controller-84d6567774-b47sz\" (UID: \"70e35241-b7b3-4693-b7bb-a439ee947d03\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.481632 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq47m\" (UniqueName: \"kubernetes.io/projected/2bf05124-0e4f-4262-95c6-d87ede36dcb5-kube-api-access-mq47m\") pod \"multus-admission-controller-857f4d67dd-55vjd\" (UID: \"2bf05124-0e4f-4262-95c6-d87ede36dcb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.503354 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.507324 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2j2s\" (UniqueName: \"kubernetes.io/projected/0d499682-90a9-464e-9cb5-6c5a82155311-kube-api-access-b2j2s\") pod \"migrator-59844c95c7-rmhjj\" (UID: \"0d499682-90a9-464e-9cb5-6c5a82155311\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.507896 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.524027 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.524450 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.528477 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk7dd\" (UniqueName: \"kubernetes.io/projected/8d60f83e-1c5e-4839-89c4-fb49928e91a8-kube-api-access-lk7dd\") pod \"dns-operator-744455d44c-rnqdz\" (UID: \"8d60f83e-1c5e-4839-89c4-fb49928e91a8\") " pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.528700 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.542611 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7d76\" (UniqueName: \"kubernetes.io/projected/195af045-05d5-4240-8115-5772fb13d082-kube-api-access-b7d76\") pod \"console-operator-58897d9998-k49s4\" (UID: \"195af045-05d5-4240-8115-5772fb13d082\") " pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.567406 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zmkp\" (UniqueName: \"kubernetes.io/projected/e49fad7d-a39a-4270-9d25-50d381408838-kube-api-access-9zmkp\") pod \"apiserver-7bbb656c7d-fdw9w\" (UID: \"e49fad7d-a39a-4270-9d25-50d381408838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.586185 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.586476 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpgvq\" (UniqueName: \"kubernetes.io/projected/c019c6e3-3a0d-4838-a28d-2392d45c52b7-kube-api-access-cpgvq\") pod \"kube-storage-version-migrator-operator-b67b599dd-r2lc6\" (UID: \"c019c6e3-3a0d-4838-a28d-2392d45c52b7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.598931 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.599171 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjfrq\" (UniqueName: \"kubernetes.io/projected/ae8e0e95-b76c-4353-b5af-2fd044fa9be3-kube-api-access-jjfrq\") pod \"ingress-operator-5b745b69d9-vm8sf\" (UID: \"ae8e0e95-b76c-4353-b5af-2fd044fa9be3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.599423 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" event={"ID":"2446a52a-a882-4ebd-972d-eb61d2e6c086","Type":"ContainerStarted","Data":"7dddfc2f8d26dae71d64f71b152e6a03ff69483f0066f51cc0d7184c89017258"} Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.599458 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" event={"ID":"2446a52a-a882-4ebd-972d-eb61d2e6c086","Type":"ContainerStarted","Data":"c4de0e734d9b5f5e27616d4b5fef21ed4d797dac1b5068bc3b7373949f0eba35"} Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.619679 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-56tjb"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.620699 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621336 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/773e3cb5-284d-448f-b025-65375eaa41c8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621358 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-tls\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621389 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/773e3cb5-284d-448f-b025-65375eaa41c8-proxy-tls\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621405 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb6x7\" (UniqueName: \"kubernetes.io/projected/773e3cb5-284d-448f-b025-65375eaa41c8-kube-api-access-cb6x7\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621513 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/773e3cb5-284d-448f-b025-65375eaa41c8-images\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621533 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-certificates\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621674 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621696 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5224ecc-afd0-4226-acd8-cb2c4197639c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621723 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-bound-sa-token\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621740 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhvwl\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-kube-api-access-dhvwl\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621759 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5224ecc-afd0-4226-acd8-cb2c4197639c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.621774 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-trusted-ca\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: E1124 14:21:55.622062 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.122050418 +0000 UTC m=+153.238690895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.636078 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.636181 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.636954 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.652808 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gzxq"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.662996 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.685612 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.695915 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w4gtt" event={"ID":"338efc62-c0d6-4589-a6a8-dd783e0f08ac","Type":"ContainerStarted","Data":"02aee72e896d79917da2676f3c5d40317b558aacb82db8852ca3731c1464d40b"} Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.697054 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" event={"ID":"c9acfdd4-def9-4f63-aec3-2f739beae15d","Type":"ContainerStarted","Data":"3fffed5e0793dd81588fb8bbd9d3ad163e14cfbf14c88ad27241b0b49809c3b4"} Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.697477 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.723096 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.724023 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:21:55 crc kubenswrapper[4822]: E1124 14:21:55.724826 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.224795117 +0000 UTC m=+153.341435604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.724889 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.724924 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5224ecc-afd0-4226-acd8-cb2c4197639c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725011 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-bound-sa-token\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725045 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhvwl\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-kube-api-access-dhvwl\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725082 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5224ecc-afd0-4226-acd8-cb2c4197639c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725113 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-trusted-ca\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725241 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/773e3cb5-284d-448f-b025-65375eaa41c8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725282 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-tls\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725401 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/773e3cb5-284d-448f-b025-65375eaa41c8-proxy-tls\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725441 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb6x7\" (UniqueName: \"kubernetes.io/projected/773e3cb5-284d-448f-b025-65375eaa41c8-kube-api-access-cb6x7\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725468 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/773e3cb5-284d-448f-b025-65375eaa41c8-images\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.725496 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-certificates\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.727123 4822 patch_prober.go:28] interesting pod/downloads-7954f5f757-p9tw4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" start-of-body= Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.727172 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-p9tw4" podUID="911c455a-0bf1-4855-956f-34a7d3f2f080" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.728530 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5224ecc-afd0-4226-acd8-cb2c4197639c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.728594 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/773e3cb5-284d-448f-b025-65375eaa41c8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: E1124 14:21:55.728722 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.228710212 +0000 UTC m=+153.345350689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.730515 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-trusted-ca\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.736184 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/773e3cb5-284d-448f-b025-65375eaa41c8-images\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.737022 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5224ecc-afd0-4226-acd8-cb2c4197639c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.737350 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-tls\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.748266 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/773e3cb5-284d-448f-b025-65375eaa41c8-proxy-tls\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.755769 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-certificates\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.767165 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-p9tw4" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.767217 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.767233 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5v7wp"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.767243 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-p9tw4" event={"ID":"911c455a-0bf1-4855-956f-34a7d3f2f080","Type":"ContainerStarted","Data":"627c82d09bd6c9722e93ee80d668ac3b67eab5e3aaf0b4f6f48b5df97e990c5f"} Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.767266 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5fdmt"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.769311 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ctks4"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.775454 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.781973 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" event={"ID":"72315396-ab30-4736-8eb9-f735ef0e7f97","Type":"ContainerStarted","Data":"13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f"} Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.782039 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" event={"ID":"72315396-ab30-4736-8eb9-f735ef0e7f97","Type":"ContainerStarted","Data":"3a9098f789c9bc1afb342643ff741d594840f520c39b017051d9b747fd8dbcdb"} Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.782992 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.795120 4822 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-hl4bh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.795174 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" podUID="72315396-ab30-4736-8eb9-f735ef0e7f97" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.804027 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhvwl\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-kube-api-access-dhvwl\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.818640 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-bound-sa-token\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.822459 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb6x7\" (UniqueName: \"kubernetes.io/projected/773e3cb5-284d-448f-b025-65375eaa41c8-kube-api-access-cb6x7\") pod \"machine-config-operator-74547568cd-7tt8s\" (UID: \"773e3cb5-284d-448f-b025-65375eaa41c8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827007 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827180 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3212b9b-7c39-4243-b7c6-afd009c4f6df-cert\") pod \"ingress-canary-qldnw\" (UID: \"f3212b9b-7c39-4243-b7c6-afd009c4f6df\") " pod="openshift-ingress-canary/ingress-canary-qldnw" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827249 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/43181ae3-6a15-4ba4-b1e0-850ed07bb1c5-signing-key\") pod \"service-ca-9c57cc56f-v69sk\" (UID: \"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5\") " pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827286 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f7ff89a-1139-432a-9890-e1fe54077965-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hm5r8\" (UID: \"8f7ff89a-1139-432a-9890-e1fe54077965\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827351 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-socket-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827386 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/81e0e571-ef24-47df-a054-54226fcfe214-tmpfs\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827427 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-registration-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827451 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-95knp\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827485 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3486a3b5-cba7-4d68-b377-310719458383-config\") pod \"service-ca-operator-777779d784-hkt6b\" (UID: \"3486a3b5-cba7-4d68-b377-310719458383\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827514 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d2hk\" (UniqueName: \"kubernetes.io/projected/498b2ff4-3940-4b34-89d2-fa7d4dc96a5e-kube-api-access-7d2hk\") pod \"dns-default-dq7pg\" (UID: \"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e\") " pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827563 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-mountpoint-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827655 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-95knp\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827705 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/979147c0-9b48-4a0f-9506-00b31ca99603-config-volume\") pod \"collect-profiles-29399895-j6qx5\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827740 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf25k\" (UniqueName: \"kubernetes.io/projected/8f7ff89a-1139-432a-9890-e1fe54077965-kube-api-access-nf25k\") pod \"olm-operator-6b444d44fb-hm5r8\" (UID: \"8f7ff89a-1139-432a-9890-e1fe54077965\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827755 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3486a3b5-cba7-4d68-b377-310719458383-serving-cert\") pod \"service-ca-operator-777779d784-hkt6b\" (UID: \"3486a3b5-cba7-4d68-b377-310719458383\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827783 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bjj2\" (UniqueName: \"kubernetes.io/projected/43181ae3-6a15-4ba4-b1e0-850ed07bb1c5-kube-api-access-7bjj2\") pod \"service-ca-9c57cc56f-v69sk\" (UID: \"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5\") " pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827838 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/762b2e27-8597-415c-b9fc-5b418baa96be-certs\") pod \"machine-config-server-k6s8b\" (UID: \"762b2e27-8597-415c-b9fc-5b418baa96be\") " pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827867 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p6rv\" (UniqueName: \"kubernetes.io/projected/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-kube-api-access-8p6rv\") pod \"marketplace-operator-79b997595-95knp\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827898 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77w4b\" (UniqueName: \"kubernetes.io/projected/762b2e27-8597-415c-b9fc-5b418baa96be-kube-api-access-77w4b\") pod \"machine-config-server-k6s8b\" (UID: \"762b2e27-8597-415c-b9fc-5b418baa96be\") " pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.827976 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f7ff89a-1139-432a-9890-e1fe54077965-srv-cert\") pod \"olm-operator-6b444d44fb-hm5r8\" (UID: \"8f7ff89a-1139-432a-9890-e1fe54077965\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828099 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81e0e571-ef24-47df-a054-54226fcfe214-webhook-cert\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828138 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7qmh\" (UniqueName: \"kubernetes.io/projected/81e0e571-ef24-47df-a054-54226fcfe214-kube-api-access-z7qmh\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828171 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlf5x\" (UniqueName: \"kubernetes.io/projected/c9c4bffa-f289-4963-879e-1d54fc224dd3-kube-api-access-xlf5x\") pod \"catalog-operator-68c6474976-dlp4b\" (UID: \"c9c4bffa-f289-4963-879e-1d54fc224dd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828237 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/498b2ff4-3940-4b34-89d2-fa7d4dc96a5e-metrics-tls\") pod \"dns-default-dq7pg\" (UID: \"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e\") " pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828279 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flvwt\" (UniqueName: \"kubernetes.io/projected/f3212b9b-7c39-4243-b7c6-afd009c4f6df-kube-api-access-flvwt\") pod \"ingress-canary-qldnw\" (UID: \"f3212b9b-7c39-4243-b7c6-afd009c4f6df\") " pod="openshift-ingress-canary/ingress-canary-qldnw" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828297 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/762b2e27-8597-415c-b9fc-5b418baa96be-node-bootstrap-token\") pod \"machine-config-server-k6s8b\" (UID: \"762b2e27-8597-415c-b9fc-5b418baa96be\") " pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828592 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81e0e571-ef24-47df-a054-54226fcfe214-apiservice-cert\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828616 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvbzw\" (UniqueName: \"kubernetes.io/projected/0442f63a-2e24-493c-9049-19bbfc0ee8ef-kube-api-access-wvbzw\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828643 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsk9q\" (UniqueName: \"kubernetes.io/projected/979147c0-9b48-4a0f-9506-00b31ca99603-kube-api-access-vsk9q\") pod \"collect-profiles-29399895-j6qx5\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828686 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-plugins-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828765 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/979147c0-9b48-4a0f-9506-00b31ca99603-secret-volume\") pod \"collect-profiles-29399895-j6qx5\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.828819 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd6cf\" (UniqueName: \"kubernetes.io/projected/3486a3b5-cba7-4d68-b377-310719458383-kube-api-access-wd6cf\") pod \"service-ca-operator-777779d784-hkt6b\" (UID: \"3486a3b5-cba7-4d68-b377-310719458383\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:55 crc kubenswrapper[4822]: E1124 14:21:55.830859 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.330840463 +0000 UTC m=+153.447480940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.831247 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/43181ae3-6a15-4ba4-b1e0-850ed07bb1c5-signing-cabundle\") pod \"service-ca-9c57cc56f-v69sk\" (UID: \"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5\") " pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.831296 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c9c4bffa-f289-4963-879e-1d54fc224dd3-srv-cert\") pod \"catalog-operator-68c6474976-dlp4b\" (UID: \"c9c4bffa-f289-4963-879e-1d54fc224dd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.831317 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/498b2ff4-3940-4b34-89d2-fa7d4dc96a5e-config-volume\") pod \"dns-default-dq7pg\" (UID: \"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e\") " pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.834479 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c9c4bffa-f289-4963-879e-1d54fc224dd3-profile-collector-cert\") pod \"catalog-operator-68c6474976-dlp4b\" (UID: \"c9c4bffa-f289-4963-879e-1d54fc224dd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.834601 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-csi-data-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.842772 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.871900 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9"] Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.936609 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f7ff89a-1139-432a-9890-e1fe54077965-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hm5r8\" (UID: \"8f7ff89a-1139-432a-9890-e1fe54077965\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937004 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-socket-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937039 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/81e0e571-ef24-47df-a054-54226fcfe214-tmpfs\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937082 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-registration-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937108 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-95knp\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937135 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3486a3b5-cba7-4d68-b377-310719458383-config\") pod \"service-ca-operator-777779d784-hkt6b\" (UID: \"3486a3b5-cba7-4d68-b377-310719458383\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937158 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d2hk\" (UniqueName: \"kubernetes.io/projected/498b2ff4-3940-4b34-89d2-fa7d4dc96a5e-kube-api-access-7d2hk\") pod \"dns-default-dq7pg\" (UID: \"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e\") " pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937184 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-mountpoint-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937231 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-95knp\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937274 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/979147c0-9b48-4a0f-9506-00b31ca99603-config-volume\") pod \"collect-profiles-29399895-j6qx5\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937357 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-socket-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937461 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf25k\" (UniqueName: \"kubernetes.io/projected/8f7ff89a-1139-432a-9890-e1fe54077965-kube-api-access-nf25k\") pod \"olm-operator-6b444d44fb-hm5r8\" (UID: \"8f7ff89a-1139-432a-9890-e1fe54077965\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937487 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-mountpoint-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937508 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3486a3b5-cba7-4d68-b377-310719458383-serving-cert\") pod \"service-ca-operator-777779d784-hkt6b\" (UID: \"3486a3b5-cba7-4d68-b377-310719458383\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937533 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bjj2\" (UniqueName: \"kubernetes.io/projected/43181ae3-6a15-4ba4-b1e0-850ed07bb1c5-kube-api-access-7bjj2\") pod \"service-ca-9c57cc56f-v69sk\" (UID: \"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5\") " pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937560 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/762b2e27-8597-415c-b9fc-5b418baa96be-certs\") pod \"machine-config-server-k6s8b\" (UID: \"762b2e27-8597-415c-b9fc-5b418baa96be\") " pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937584 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p6rv\" (UniqueName: \"kubernetes.io/projected/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-kube-api-access-8p6rv\") pod \"marketplace-operator-79b997595-95knp\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937607 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77w4b\" (UniqueName: \"kubernetes.io/projected/762b2e27-8597-415c-b9fc-5b418baa96be-kube-api-access-77w4b\") pod \"machine-config-server-k6s8b\" (UID: \"762b2e27-8597-415c-b9fc-5b418baa96be\") " pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937644 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f7ff89a-1139-432a-9890-e1fe54077965-srv-cert\") pod \"olm-operator-6b444d44fb-hm5r8\" (UID: \"8f7ff89a-1139-432a-9890-e1fe54077965\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937673 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81e0e571-ef24-47df-a054-54226fcfe214-webhook-cert\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937699 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7qmh\" (UniqueName: \"kubernetes.io/projected/81e0e571-ef24-47df-a054-54226fcfe214-kube-api-access-z7qmh\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937730 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlf5x\" (UniqueName: \"kubernetes.io/projected/c9c4bffa-f289-4963-879e-1d54fc224dd3-kube-api-access-xlf5x\") pod \"catalog-operator-68c6474976-dlp4b\" (UID: \"c9c4bffa-f289-4963-879e-1d54fc224dd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937757 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/498b2ff4-3940-4b34-89d2-fa7d4dc96a5e-metrics-tls\") pod \"dns-default-dq7pg\" (UID: \"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e\") " pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937785 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flvwt\" (UniqueName: \"kubernetes.io/projected/f3212b9b-7c39-4243-b7c6-afd009c4f6df-kube-api-access-flvwt\") pod \"ingress-canary-qldnw\" (UID: \"f3212b9b-7c39-4243-b7c6-afd009c4f6df\") " pod="openshift-ingress-canary/ingress-canary-qldnw" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937808 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/762b2e27-8597-415c-b9fc-5b418baa96be-node-bootstrap-token\") pod \"machine-config-server-k6s8b\" (UID: \"762b2e27-8597-415c-b9fc-5b418baa96be\") " pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937836 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81e0e571-ef24-47df-a054-54226fcfe214-apiservice-cert\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937866 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvbzw\" (UniqueName: \"kubernetes.io/projected/0442f63a-2e24-493c-9049-19bbfc0ee8ef-kube-api-access-wvbzw\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937901 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937929 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-plugins-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937955 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsk9q\" (UniqueName: \"kubernetes.io/projected/979147c0-9b48-4a0f-9506-00b31ca99603-kube-api-access-vsk9q\") pod \"collect-profiles-29399895-j6qx5\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937980 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3486a3b5-cba7-4d68-b377-310719458383-config\") pod \"service-ca-operator-777779d784-hkt6b\" (UID: \"3486a3b5-cba7-4d68-b377-310719458383\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.937988 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/979147c0-9b48-4a0f-9506-00b31ca99603-secret-volume\") pod \"collect-profiles-29399895-j6qx5\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938027 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd6cf\" (UniqueName: \"kubernetes.io/projected/3486a3b5-cba7-4d68-b377-310719458383-kube-api-access-wd6cf\") pod \"service-ca-operator-777779d784-hkt6b\" (UID: \"3486a3b5-cba7-4d68-b377-310719458383\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938054 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/43181ae3-6a15-4ba4-b1e0-850ed07bb1c5-signing-cabundle\") pod \"service-ca-9c57cc56f-v69sk\" (UID: \"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5\") " pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938089 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c9c4bffa-f289-4963-879e-1d54fc224dd3-srv-cert\") pod \"catalog-operator-68c6474976-dlp4b\" (UID: \"c9c4bffa-f289-4963-879e-1d54fc224dd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938107 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/498b2ff4-3940-4b34-89d2-fa7d4dc96a5e-config-volume\") pod \"dns-default-dq7pg\" (UID: \"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e\") " pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938133 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c9c4bffa-f289-4963-879e-1d54fc224dd3-profile-collector-cert\") pod \"catalog-operator-68c6474976-dlp4b\" (UID: \"c9c4bffa-f289-4963-879e-1d54fc224dd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938166 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-csi-data-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938191 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3212b9b-7c39-4243-b7c6-afd009c4f6df-cert\") pod \"ingress-canary-qldnw\" (UID: \"f3212b9b-7c39-4243-b7c6-afd009c4f6df\") " pod="openshift-ingress-canary/ingress-canary-qldnw" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938257 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/43181ae3-6a15-4ba4-b1e0-850ed07bb1c5-signing-key\") pod \"service-ca-9c57cc56f-v69sk\" (UID: \"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5\") " pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938361 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/81e0e571-ef24-47df-a054-54226fcfe214-tmpfs\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938422 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/979147c0-9b48-4a0f-9506-00b31ca99603-config-volume\") pod \"collect-profiles-29399895-j6qx5\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.938477 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-registration-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.939030 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-95knp\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:55 crc kubenswrapper[4822]: E1124 14:21:55.939470 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.439454711 +0000 UTC m=+153.556095278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.939536 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-plugins-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.939554 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/43181ae3-6a15-4ba4-b1e0-850ed07bb1c5-signing-cabundle\") pod \"service-ca-9c57cc56f-v69sk\" (UID: \"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5\") " pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.940962 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0442f63a-2e24-493c-9049-19bbfc0ee8ef-csi-data-dir\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.940993 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/498b2ff4-3940-4b34-89d2-fa7d4dc96a5e-config-volume\") pod \"dns-default-dq7pg\" (UID: \"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e\") " pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.945603 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f7ff89a-1139-432a-9890-e1fe54077965-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hm5r8\" (UID: \"8f7ff89a-1139-432a-9890-e1fe54077965\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.946369 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/43181ae3-6a15-4ba4-b1e0-850ed07bb1c5-signing-key\") pod \"service-ca-9c57cc56f-v69sk\" (UID: \"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5\") " pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.950431 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/498b2ff4-3940-4b34-89d2-fa7d4dc96a5e-metrics-tls\") pod \"dns-default-dq7pg\" (UID: \"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e\") " pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.952329 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c9c4bffa-f289-4963-879e-1d54fc224dd3-profile-collector-cert\") pod \"catalog-operator-68c6474976-dlp4b\" (UID: \"c9c4bffa-f289-4963-879e-1d54fc224dd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.952387 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/979147c0-9b48-4a0f-9506-00b31ca99603-secret-volume\") pod \"collect-profiles-29399895-j6qx5\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.952549 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/762b2e27-8597-415c-b9fc-5b418baa96be-certs\") pod \"machine-config-server-k6s8b\" (UID: \"762b2e27-8597-415c-b9fc-5b418baa96be\") " pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.952855 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.952866 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-95knp\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.953976 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3212b9b-7c39-4243-b7c6-afd009c4f6df-cert\") pod \"ingress-canary-qldnw\" (UID: \"f3212b9b-7c39-4243-b7c6-afd009c4f6df\") " pod="openshift-ingress-canary/ingress-canary-qldnw" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.954132 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3486a3b5-cba7-4d68-b377-310719458383-serving-cert\") pod \"service-ca-operator-777779d784-hkt6b\" (UID: \"3486a3b5-cba7-4d68-b377-310719458383\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.959625 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c9c4bffa-f289-4963-879e-1d54fc224dd3-srv-cert\") pod \"catalog-operator-68c6474976-dlp4b\" (UID: \"c9c4bffa-f289-4963-879e-1d54fc224dd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.964229 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f7ff89a-1139-432a-9890-e1fe54077965-srv-cert\") pod \"olm-operator-6b444d44fb-hm5r8\" (UID: \"8f7ff89a-1139-432a-9890-e1fe54077965\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.964995 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d2hk\" (UniqueName: \"kubernetes.io/projected/498b2ff4-3940-4b34-89d2-fa7d4dc96a5e-kube-api-access-7d2hk\") pod \"dns-default-dq7pg\" (UID: \"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e\") " pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.967589 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81e0e571-ef24-47df-a054-54226fcfe214-webhook-cert\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.979363 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/762b2e27-8597-415c-b9fc-5b418baa96be-node-bootstrap-token\") pod \"machine-config-server-k6s8b\" (UID: \"762b2e27-8597-415c-b9fc-5b418baa96be\") " pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:55 crc kubenswrapper[4822]: I1124 14:21:55.979911 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81e0e571-ef24-47df-a054-54226fcfe214-apiservice-cert\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.002957 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf25k\" (UniqueName: \"kubernetes.io/projected/8f7ff89a-1139-432a-9890-e1fe54077965-kube-api-access-nf25k\") pod \"olm-operator-6b444d44fb-hm5r8\" (UID: \"8f7ff89a-1139-432a-9890-e1fe54077965\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.005319 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.011009 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvbzw\" (UniqueName: \"kubernetes.io/projected/0442f63a-2e24-493c-9049-19bbfc0ee8ef-kube-api-access-wvbzw\") pod \"csi-hostpathplugin-6rhjd\" (UID: \"0442f63a-2e24-493c-9049-19bbfc0ee8ef\") " pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.033153 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsk9q\" (UniqueName: \"kubernetes.io/projected/979147c0-9b48-4a0f-9506-00b31ca99603-kube-api-access-vsk9q\") pod \"collect-profiles-29399895-j6qx5\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.038737 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.039101 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.539081842 +0000 UTC m=+153.655722319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.039732 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-k49s4"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.056335 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p6rv\" (UniqueName: \"kubernetes.io/projected/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-kube-api-access-8p6rv\") pod \"marketplace-operator-79b997595-95knp\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.057031 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.060273 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.086261 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.097405 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bjj2\" (UniqueName: \"kubernetes.io/projected/43181ae3-6a15-4ba4-b1e0-850ed07bb1c5-kube-api-access-7bjj2\") pod \"service-ca-9c57cc56f-v69sk\" (UID: \"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5\") " pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.098052 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7qmh\" (UniqueName: \"kubernetes.io/projected/81e0e571-ef24-47df-a054-54226fcfe214-kube-api-access-z7qmh\") pod \"packageserver-d55dfcdfc-ftv2h\" (UID: \"81e0e571-ef24-47df-a054-54226fcfe214\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.104494 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dq7pg" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.110990 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlf5x\" (UniqueName: \"kubernetes.io/projected/c9c4bffa-f289-4963-879e-1d54fc224dd3-kube-api-access-xlf5x\") pod \"catalog-operator-68c6474976-dlp4b\" (UID: \"c9c4bffa-f289-4963-879e-1d54fc224dd3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.128919 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.140471 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.141234 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.641214052 +0000 UTC m=+153.757854529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.150151 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77w4b\" (UniqueName: \"kubernetes.io/projected/762b2e27-8597-415c-b9fc-5b418baa96be-kube-api-access-77w4b\") pod \"machine-config-server-k6s8b\" (UID: \"762b2e27-8597-415c-b9fc-5b418baa96be\") " pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.157271 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.157502 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8bc7n"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.163065 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flvwt\" (UniqueName: \"kubernetes.io/projected/f3212b9b-7c39-4243-b7c6-afd009c4f6df-kube-api-access-flvwt\") pod \"ingress-canary-qldnw\" (UID: \"f3212b9b-7c39-4243-b7c6-afd009c4f6df\") " pod="openshift-ingress-canary/ingress-canary-qldnw" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.163197 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.184688 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd6cf\" (UniqueName: \"kubernetes.io/projected/3486a3b5-cba7-4d68-b377-310719458383-kube-api-access-wd6cf\") pod \"service-ca-operator-777779d784-hkt6b\" (UID: \"3486a3b5-cba7-4d68-b377-310719458383\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:56 crc kubenswrapper[4822]: W1124 14:21:56.189931 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod195af045_05d5_4240_8115_5772fb13d082.slice/crio-21471860bf0960ab61dd1da7ff2629cf251428a2140db38aabc0e0b5c5bc8429 WatchSource:0}: Error finding container 21471860bf0960ab61dd1da7ff2629cf251428a2140db38aabc0e0b5c5bc8429: Status 404 returned error can't find the container with id 21471860bf0960ab61dd1da7ff2629cf251428a2140db38aabc0e0b5c5bc8429 Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.246197 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.247572 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.747554457 +0000 UTC m=+153.864194934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.330690 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.337163 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-p9tw4" podStartSLOduration=130.337142816 podStartE2EDuration="2m10.337142816s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:56.317839167 +0000 UTC m=+153.434479664" watchObservedRunningTime="2025-11-24 14:21:56.337142816 +0000 UTC m=+153.453783293" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.337700 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rnqdz"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.338525 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.346865 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.353328 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.353833 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.853807899 +0000 UTC m=+153.970448376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.364372 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.372417 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.396240 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-k6s8b" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.432451 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.434611 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qldnw" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.454314 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.455227 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:56.955171325 +0000 UTC m=+154.071811802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:56 crc kubenswrapper[4822]: W1124 14:21:56.506750 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d60f83e_1c5e_4839_89c4_fb49928e91a8.slice/crio-63c88b8de98294423273437b38cdb096fcbab1ff41baaa2ff220b04d8c0224de WatchSource:0}: Error finding container 63c88b8de98294423273437b38cdb096fcbab1ff41baaa2ff220b04d8c0224de: Status 404 returned error can't find the container with id 63c88b8de98294423273437b38cdb096fcbab1ff41baaa2ff220b04d8c0224de Nov 24 14:21:56 crc kubenswrapper[4822]: W1124 14:21:56.507280 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d499682_90a9_464e_9cb5_6c5a82155311.slice/crio-13f7f57235e55306166473d04797ebf3eef8f75c8b2d4511183a22319463c09a WatchSource:0}: Error finding container 13f7f57235e55306166473d04797ebf3eef8f75c8b2d4511183a22319463c09a: Status 404 returned error can't find the container with id 13f7f57235e55306166473d04797ebf3eef8f75c8b2d4511183a22319463c09a Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.559388 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.559820 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.059803726 +0000 UTC m=+154.176444203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.567693 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.628176 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.644082 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-55vjd"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.669088 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.700382 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.169734685 +0000 UTC m=+154.286375162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.701700 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf"] Nov 24 14:21:56 crc kubenswrapper[4822]: W1124 14:21:56.717288 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70e35241_b7b3_4693_b7bb_a439ee947d03.slice/crio-9fe79024444775b8dbfee51ccd11435a4bc882305fe2e8d87a42e820c6cb05ca WatchSource:0}: Error finding container 9fe79024444775b8dbfee51ccd11435a4bc882305fe2e8d87a42e820c6cb05ca: Status 404 returned error can't find the container with id 9fe79024444775b8dbfee51ccd11435a4bc882305fe2e8d87a42e820c6cb05ca Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.739486 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.778454 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.778882 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.27886941 +0000 UTC m=+154.395509887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.818095 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" event={"ID":"6b8514e2-4241-4916-b843-d62897d4af2c","Type":"ContainerStarted","Data":"800ad43b0faca900632589fff4720f65d3ac6e627163d67a41b04a60ef1f72bb"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.818506 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" event={"ID":"6b8514e2-4241-4916-b843-d62897d4af2c","Type":"ContainerStarted","Data":"73976d58139b4ca589a7a02a8f05eaed900dc4475dc4788e83984dc2a13a9d35"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.855102 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" event={"ID":"397bb803-fafb-4fca-ae9f-e373d413958f","Type":"ContainerStarted","Data":"65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.855154 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" event={"ID":"397bb803-fafb-4fca-ae9f-e373d413958f","Type":"ContainerStarted","Data":"0b83989637e7062aab70e259f63105b0a72baf106f0415824ce1b59d3e404da3"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.855895 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.873869 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" event={"ID":"8a1a53d5-5871-4b71-8d79-f496da018491","Type":"ContainerStarted","Data":"f5027ccb94f269490e569ce65bba50dac8fc77f055ed4c2e620bbc367e51a59a"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.879229 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.880244 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.380227586 +0000 UTC m=+154.496868063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.883874 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.896544 4822 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-q7wf8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.896602 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" podUID="397bb803-fafb-4fca-ae9f-e373d413958f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.924099 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.950575 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" event={"ID":"e83d919d-2e04-43b4-9f0c-40aba2d18772","Type":"ContainerStarted","Data":"35aca5258e4e19546553d6bddc5eef907c16c23b416c99702b1621a23ad76cb9"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.950647 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" event={"ID":"e83d919d-2e04-43b4-9f0c-40aba2d18772","Type":"ContainerStarted","Data":"8c68f012a4fbf148aa6497df69570fe48beb01ec18155b0f1ac00c1b424792c6"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.961247 4822 generic.go:334] "Generic (PLEG): container finished" podID="90001d98-f4e4-42f3-8f78-453f244a6286" containerID="e4181a1845d6fbd620d7a0771e9ed1cdeb17ab0260fe787fdf84a05a59a37be8" exitCode=0 Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.961384 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" event={"ID":"90001d98-f4e4-42f3-8f78-453f244a6286","Type":"ContainerDied","Data":"e4181a1845d6fbd620d7a0771e9ed1cdeb17ab0260fe787fdf84a05a59a37be8"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.961423 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" event={"ID":"90001d98-f4e4-42f3-8f78-453f244a6286","Type":"ContainerStarted","Data":"07183e09b332bf6045697e7604ef86dbfeb1a17c1c24c78e87d5c9337af7882b"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.963887 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8"] Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.965033 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" event={"ID":"15f981aa-3735-40f5-bc13-2cfad43ab9d8","Type":"ContainerStarted","Data":"6b019ba51ad39c04e4272aadd903350612fd8ccd34535225c8a2dd4f10c729ea"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.969840 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" event={"ID":"2446a52a-a882-4ebd-972d-eb61d2e6c086","Type":"ContainerStarted","Data":"6f7bfb9a995a01b1a49b1b9020aa4f8d58abf08a5951e94c8add9c4e5d5f5246"} Nov 24 14:21:56 crc kubenswrapper[4822]: I1124 14:21:56.982896 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:56 crc kubenswrapper[4822]: E1124 14:21:56.989620 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.489594778 +0000 UTC m=+154.606235445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.012376 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6rhjd"] Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.016134 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-p9tw4" event={"ID":"911c455a-0bf1-4855-956f-34a7d3f2f080","Type":"ContainerStarted","Data":"9027d130da55083ad6dd6226ff3c73a118f2b1f3581021b6238ec9889a49877d"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.020695 4822 patch_prober.go:28] interesting pod/downloads-7954f5f757-p9tw4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" start-of-body= Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.020862 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-p9tw4" podUID="911c455a-0bf1-4855-956f-34a7d3f2f080" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.087141 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.087402 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.587365259 +0000 UTC m=+154.704005736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.087496 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.088548 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.588532167 +0000 UTC m=+154.705172644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.101051 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dq7pg"] Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.110138 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" event={"ID":"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12","Type":"ContainerStarted","Data":"3afcf582cf01e352bb84903b1c9e43a30ffa4c135b4815b6fab76035a833a3ea"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.110183 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" event={"ID":"3e9606ec-b2ea-4f43-b19c-b4d0a0e3bc12","Type":"ContainerStarted","Data":"fbeab1645cf2e9939cfaffa1ddfd706c431149291fd39bda481e6dd799b9e619"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.115564 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" event={"ID":"8593f46e-611a-46a9-9644-aed898afd907","Type":"ContainerStarted","Data":"ee223c7eb1d2a97eef2beba724e2a88896d377a89fe79dae56b38786846447c2"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.133307 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" event={"ID":"c9acfdd4-def9-4f63-aec3-2f739beae15d","Type":"ContainerStarted","Data":"9a4c20a74a5f734098b46978399d7d7a996307ce03ed9b45121a909d8ddb4b7d"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.133363 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" event={"ID":"c9acfdd4-def9-4f63-aec3-2f739beae15d","Type":"ContainerStarted","Data":"c6f6a0293a168af33182a376c6d5efeceb1159c769506cb75e3e7f07ab60ba89"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.144227 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" event={"ID":"8d60f83e-1c5e-4839-89c4-fb49928e91a8","Type":"ContainerStarted","Data":"63c88b8de98294423273437b38cdb096fcbab1ff41baaa2ff220b04d8c0224de"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.145671 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" event={"ID":"b934905c-f8df-4e00-9926-19df23d81e61","Type":"ContainerStarted","Data":"8e954e446c45e42cf937cd0270e864ec5adfed4cf4c6d4d90563caea2d407eb2"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.194753 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" event={"ID":"02d5de9e-3cc2-46a7-8138-cca5157799e5","Type":"ContainerStarted","Data":"7a30418d735ac7f646fdbab2be096f7bbacb9954f6146f0775044ba21cdd0d63"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.195000 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.195084 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.695065538 +0000 UTC m=+154.811706015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.196042 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.200183 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.700158781 +0000 UTC m=+154.816799248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.247354 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" event={"ID":"902bd164-754e-4c87-b042-ca87c15d0263","Type":"ContainerStarted","Data":"37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.247420 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" event={"ID":"902bd164-754e-4c87-b042-ca87c15d0263","Type":"ContainerStarted","Data":"0ad157597f9cf9931ec4ba2591d5112aad1651dc0e1fadba8465e4851a4d282f"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.248837 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.260483 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w4gtt" event={"ID":"338efc62-c0d6-4589-a6a8-dd783e0f08ac","Type":"ContainerStarted","Data":"cff0bd5558970ceaf1890aaa2d9fefa94c82d85e13ba23a53d2361c530d7ca7a"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.261471 4822 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7gzxq container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" start-of-body= Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.261541 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" podUID="902bd164-754e-4c87-b042-ca87c15d0263" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.261842 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w"] Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.265607 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj" event={"ID":"0d499682-90a9-464e-9cb5-6c5a82155311","Type":"ContainerStarted","Data":"13f7f57235e55306166473d04797ebf3eef8f75c8b2d4511183a22319463c09a"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.287555 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" event={"ID":"985f7b3a-f9d1-44cc-9132-6e24bc94e8f3","Type":"ContainerStarted","Data":"8cc0065cc32b9f5dce9d2e260ecdf182f692357f6c7451558d54936f870bdecb"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.298388 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.298836 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.79881436 +0000 UTC m=+154.915454837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.304067 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" event={"ID":"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b","Type":"ContainerStarted","Data":"6aa36adc85c6b09468fcd25d8f59f1801f9344d524e894b88b3eb00a66508020"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.304143 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" event={"ID":"58ae8ccc-42ce-4e8f-a1fc-bbe53cedf25b","Type":"ContainerStarted","Data":"06fdf5d5fe543dade3e660b5576280353ef7aec67eee520cef38ce313c88f7a5"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.352563 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.352893 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" podStartSLOduration=131.350898788 podStartE2EDuration="2m11.350898788s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:57.286068062 +0000 UTC m=+154.402708549" watchObservedRunningTime="2025-11-24 14:21:57.350898788 +0000 UTC m=+154.467539265" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.379580 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" event={"ID":"d58f06c7-b1a1-48ae-8b56-78b7f08970f5","Type":"ContainerStarted","Data":"021e06e6223c8512836e18227b3e3cb9d5283c28ae72510c013e7052de992f22"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.379656 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" event={"ID":"d58f06c7-b1a1-48ae-8b56-78b7f08970f5","Type":"ContainerStarted","Data":"383dedeb80ee1e36dd7a242291e607c97d4c3060b39affa775580d5883954204"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.404378 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.409471 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:57.909452033 +0000 UTC m=+155.026092510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.412460 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5v7wp" event={"ID":"6153af6d-c527-485f-8a0e-0cfecfb43d5b","Type":"ContainerStarted","Data":"b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.412514 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5v7wp" event={"ID":"6153af6d-c527-485f-8a0e-0cfecfb43d5b","Type":"ContainerStarted","Data":"9630f913db10497b21d6a0ef428e8ca730501d9c2a427ec53e0e1299d4ef37a5"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.465609 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" event={"ID":"70e35241-b7b3-4693-b7bb-a439ee947d03","Type":"ContainerStarted","Data":"9fe79024444775b8dbfee51ccd11435a4bc882305fe2e8d87a42e820c6cb05ca"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.478797 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-k49s4" event={"ID":"195af045-05d5-4240-8115-5772fb13d082","Type":"ContainerStarted","Data":"21471860bf0960ab61dd1da7ff2629cf251428a2140db38aabc0e0b5c5bc8429"} Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.479549 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.506464 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.508669 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.00865207 +0000 UTC m=+155.125292547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.508681 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.508780 4822 patch_prober.go:28] interesting pod/console-operator-58897d9998-k49s4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.508814 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-k49s4" podUID="195af045-05d5-4240-8115-5772fb13d082" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.608644 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.610824 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.11080355 +0000 UTC m=+155.227444027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.659900 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-k49s4" podStartSLOduration=131.659878922 podStartE2EDuration="2m11.659878922s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:57.656566886 +0000 UTC m=+154.773207373" watchObservedRunningTime="2025-11-24 14:21:57.659878922 +0000 UTC m=+154.776519399" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.709644 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.711055 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.21103542 +0000 UTC m=+155.327675907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.805711 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:21:57 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:21:57 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:21:57 crc kubenswrapper[4822]: healthz check failed Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.805784 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.815244 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.815640 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.315623829 +0000 UTC m=+155.432264306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.894066 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" podStartSLOduration=131.8940457 podStartE2EDuration="2m11.8940457s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:57.861169508 +0000 UTC m=+154.977809995" watchObservedRunningTime="2025-11-24 14:21:57.8940457 +0000 UTC m=+155.010686177" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.914801 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-qr4nb" podStartSLOduration=130.914783375 podStartE2EDuration="2m10.914783375s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:57.913742611 +0000 UTC m=+155.030383088" watchObservedRunningTime="2025-11-24 14:21:57.914783375 +0000 UTC m=+155.031423852" Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.916837 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.917177 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.41713633 +0000 UTC m=+155.533776817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:57 crc kubenswrapper[4822]: I1124 14:21:57.917648 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:57 crc kubenswrapper[4822]: E1124 14:21:57.918564 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.418554295 +0000 UTC m=+155.535194772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.019284 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-w4gtt" podStartSLOduration=131.019170937 podStartE2EDuration="2m11.019170937s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:57.960774167 +0000 UTC m=+155.077414644" watchObservedRunningTime="2025-11-24 14:21:58.019170937 +0000 UTC m=+155.135811424" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.027054 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.028292 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.528260548 +0000 UTC m=+155.644901025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.102012 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-g7jjl" podStartSLOduration=131.101977539 podStartE2EDuration="2m11.101977539s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.094066236 +0000 UTC m=+155.210706723" watchObservedRunningTime="2025-11-24 14:21:58.101977539 +0000 UTC m=+155.218618016" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.133028 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.134295 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.634255162 +0000 UTC m=+155.750895639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.156587 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-5fdmt" podStartSLOduration=132.156561506 podStartE2EDuration="2m12.156561506s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.154039975 +0000 UTC m=+155.270680472" watchObservedRunningTime="2025-11-24 14:21:58.156561506 +0000 UTC m=+155.273201983" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.226120 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vqxq9" podStartSLOduration=131.226088702 podStartE2EDuration="2m11.226088702s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.204572534 +0000 UTC m=+155.321213031" watchObservedRunningTime="2025-11-24 14:21:58.226088702 +0000 UTC m=+155.342729169" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.234402 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.234841 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.734821262 +0000 UTC m=+155.851461739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.336924 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.337342 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.837329435 +0000 UTC m=+155.953969912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.357518 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:21:58 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:21:58 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:21:58 crc kubenswrapper[4822]: healthz check failed Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.357581 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.360520 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thks6" podStartSLOduration=132.360498256 podStartE2EDuration="2m12.360498256s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.359144474 +0000 UTC m=+155.475784951" watchObservedRunningTime="2025-11-24 14:21:58.360498256 +0000 UTC m=+155.477138733" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.361506 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" podStartSLOduration=131.361499969 podStartE2EDuration="2m11.361499969s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.304914977 +0000 UTC m=+155.421555464" watchObservedRunningTime="2025-11-24 14:21:58.361499969 +0000 UTC m=+155.478140446" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.415875 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qldnw"] Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.440557 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.441032 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:58.941009675 +0000 UTC m=+156.057650152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.499495 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b"] Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.515904 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v69sk"] Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.527988 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-95knp"] Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.543069 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.543483 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.043466266 +0000 UTC m=+156.160106743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.546569 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h"] Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.546618 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" event={"ID":"e49fad7d-a39a-4270-9d25-50d381408838","Type":"ContainerStarted","Data":"f34b700430bbf3dc643168af5dd56e0b59ad23e6a021d3cbc043ac008b1526da"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.558418 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" event={"ID":"773e3cb5-284d-448f-b025-65375eaa41c8","Type":"ContainerStarted","Data":"fede70ec1268f38d5daf20da8e941992f02b4d77fcdfa067dbc79db5965de442"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.559989 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" event={"ID":"985f7b3a-f9d1-44cc-9132-6e24bc94e8f3","Type":"ContainerStarted","Data":"7f8909626f2ea9bc5236bde4702175d0276ab1cfef5b862e50c423746576f4cf"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.572894 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" event={"ID":"90001d98-f4e4-42f3-8f78-453f244a6286","Type":"ContainerStarted","Data":"170141e4a98e260657a03bb958d8a273ac3b39b73f045ddd8d1984f0dfb1015d"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.573795 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.580858 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b"] Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.593723 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" event={"ID":"0442f63a-2e24-493c-9049-19bbfc0ee8ef","Type":"ContainerStarted","Data":"4ad04ddbd8ab19882c083327a2b5de7802bbab55fcfb9ea9999d8e40ffa3fce6"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.615650 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-k6s8b" event={"ID":"762b2e27-8597-415c-b9fc-5b418baa96be","Type":"ContainerStarted","Data":"49c7e2aff2004c9d01b892535ff10983452e5bce44ce0e35c125203d1453aa27"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.615716 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-k6s8b" event={"ID":"762b2e27-8597-415c-b9fc-5b418baa96be","Type":"ContainerStarted","Data":"cbde5da3693da0b9bcc18d7657637f2e289eebccc8691789cdae908cbbcd0d5f"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.616052 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5v7wp" podStartSLOduration=132.616026589 podStartE2EDuration="2m12.616026589s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.602905039 +0000 UTC m=+155.719545536" watchObservedRunningTime="2025-11-24 14:21:58.616026589 +0000 UTC m=+155.732667066" Nov 24 14:21:58 crc kubenswrapper[4822]: W1124 14:21:58.634428 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf54974f4_d4d0_4962_baf5_d145bd8fc5f9.slice/crio-316a77d30678ea484470c8402bf7857ba3b3d927a0600dbefbaf85f88f85d532 WatchSource:0}: Error finding container 316a77d30678ea484470c8402bf7857ba3b3d927a0600dbefbaf85f88f85d532: Status 404 returned error can't find the container with id 316a77d30678ea484470c8402bf7857ba3b3d927a0600dbefbaf85f88f85d532 Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.639567 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dq7pg" event={"ID":"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e","Type":"ContainerStarted","Data":"14c1485bd14119fe1b30dc51cd582e3f568d8f1853bb2e67550bcb691bf0d720"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.644599 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" event={"ID":"8a1a53d5-5871-4b71-8d79-f496da018491","Type":"ContainerStarted","Data":"63d4f46eba397655990397ce6640aa48cafbec40d6a58e6f35f1bc6759078222"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.647806 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.648093 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.148069416 +0000 UTC m=+156.264709883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.648223 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.648522 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.148515569 +0000 UTC m=+156.265156046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.650142 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" event={"ID":"979147c0-9b48-4a0f-9506-00b31ca99603","Type":"ContainerStarted","Data":"29618aa5e61a2deb1471f441c5cae592d8f4dbc7cf6158438651040c0b6d5951"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.653489 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" event={"ID":"8f7ff89a-1139-432a-9890-e1fe54077965","Type":"ContainerStarted","Data":"3732add494a868fb95bf4d61ecbf04e4d5df5dee230113d90486f6eb192f4ebf"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.654756 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.663400 4822 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hm5r8 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.663469 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" podUID="8f7ff89a-1139-432a-9890-e1fe54077965" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.666890 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" event={"ID":"b934905c-f8df-4e00-9926-19df23d81e61","Type":"ContainerStarted","Data":"74a9890f8794a5c5fa10f1a71b0798dff95bb7b594d796b1c4af6df705bd0651"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.667770 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.671687 4822 generic.go:334] "Generic (PLEG): container finished" podID="15f981aa-3735-40f5-bc13-2cfad43ab9d8" containerID="1d1ffe8f6accb41e94e903bf1eeb4ed8155ac16c64dc836b5947c1c39a31e391" exitCode=0 Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.671767 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" event={"ID":"15f981aa-3735-40f5-bc13-2cfad43ab9d8","Type":"ContainerDied","Data":"1d1ffe8f6accb41e94e903bf1eeb4ed8155ac16c64dc836b5947c1c39a31e391"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.679276 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" event={"ID":"8593f46e-611a-46a9-9644-aed898afd907","Type":"ContainerStarted","Data":"fc8b3721d1e8884ed624b53d46e964ff4544cde3d769ba8852c63f71f5050626"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.689697 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mscgb" podStartSLOduration=132.689674238 podStartE2EDuration="2m12.689674238s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.687028683 +0000 UTC m=+155.803669170" watchObservedRunningTime="2025-11-24 14:21:58.689674238 +0000 UTC m=+155.806314715" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.696602 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" event={"ID":"c019c6e3-3a0d-4838-a28d-2392d45c52b7","Type":"ContainerStarted","Data":"0457a6b7e3915dc5fc6d192c064fe9e9ff5fc8b447dddb663f5574aa312fbea5"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.717199 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj" event={"ID":"0d499682-90a9-464e-9cb5-6c5a82155311","Type":"ContainerStarted","Data":"2482f464873d3b3b4aa5f1aa5da6ad180ef6b19bbd295a3389d071b3f2bfbe71"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.729398 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" event={"ID":"2bf05124-0e4f-4262-95c6-d87ede36dcb5","Type":"ContainerStarted","Data":"297be37f2e6180e2a4f9dac03dc22be8f9e66ba7929a07ac9ff48f63fe387e76"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.748793 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.750153 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" event={"ID":"ae8e0e95-b76c-4353-b5af-2fd044fa9be3","Type":"ContainerStarted","Data":"9564187b9dfe97f0e278be24a85d038004630fdcfb21549d1d8b2bf682db82ef"} Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.750159 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.250131503 +0000 UTC m=+156.366772020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.775501 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" event={"ID":"28e02b8d-b939-4a4c-952e-3e8365a1d124","Type":"ContainerStarted","Data":"d14c22ca620da494ffad3809cf2d907aa0f83e2449f39ed6dd23e0a10822e9f6"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.790072 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" event={"ID":"70e35241-b7b3-4693-b7bb-a439ee947d03","Type":"ContainerStarted","Data":"d36a31f935ba42f7e927cf1f1717e8c54dfe1196584b8ff3655e5dff5806cc79"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.795191 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m57q2" podStartSLOduration=131.795167206 podStartE2EDuration="2m11.795167206s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.795067442 +0000 UTC m=+155.911707919" watchObservedRunningTime="2025-11-24 14:21:58.795167206 +0000 UTC m=+155.911807683" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.806909 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-56tjb" podStartSLOduration=132.806882601 podStartE2EDuration="2m12.806882601s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.769299077 +0000 UTC m=+155.885939574" watchObservedRunningTime="2025-11-24 14:21:58.806882601 +0000 UTC m=+155.923523078" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.826021 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" event={"ID":"02d5de9e-3cc2-46a7-8138-cca5157799e5","Type":"ContainerStarted","Data":"e765676351228ba4066e21a90c9db8d8e36b3370370c53e4b1a87c01cb876068"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.852158 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.852748 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.352728609 +0000 UTC m=+156.469369086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.862468 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m56tg" podStartSLOduration=132.86243931 podStartE2EDuration="2m12.86243931s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.845228198 +0000 UTC m=+155.961868685" watchObservedRunningTime="2025-11-24 14:21:58.86243931 +0000 UTC m=+155.979079797" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.880510 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" podStartSLOduration=131.880490968 podStartE2EDuration="2m11.880490968s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.877874074 +0000 UTC m=+155.994514551" watchObservedRunningTime="2025-11-24 14:21:58.880490968 +0000 UTC m=+155.997131445" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.914951 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-k49s4" event={"ID":"195af045-05d5-4240-8115-5772fb13d082","Type":"ContainerStarted","Data":"df3b93f746b1a515261145ff75433622400a128e8d333646b8358c424e3c5b74"} Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.921038 4822 patch_prober.go:28] interesting pod/downloads-7954f5f757-p9tw4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" start-of-body= Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.921119 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-p9tw4" podUID="911c455a-0bf1-4855-956f-34a7d3f2f080" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.929587 4822 patch_prober.go:28] interesting pod/console-operator-58897d9998-k49s4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.929640 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-k49s4" podUID="195af045-05d5-4240-8115-5772fb13d082" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.943946 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.947883 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-k6s8b" podStartSLOduration=6.9478654760000005 podStartE2EDuration="6.947865476s" podCreationTimestamp="2025-11-24 14:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.946910775 +0000 UTC m=+156.063551252" watchObservedRunningTime="2025-11-24 14:21:58.947865476 +0000 UTC m=+156.064505943" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.956858 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:58 crc kubenswrapper[4822]: E1124 14:21:58.959644 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.459619541 +0000 UTC m=+156.576260018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.960258 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:21:58 crc kubenswrapper[4822]: I1124 14:21:58.984575 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" podStartSLOduration=131.98455546 podStartE2EDuration="2m11.98455546s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:58.982326138 +0000 UTC m=+156.098966625" watchObservedRunningTime="2025-11-24 14:21:58.98455546 +0000 UTC m=+156.101195937" Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.026305 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" podStartSLOduration=133.026287107 podStartE2EDuration="2m13.026287107s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:59.020745279 +0000 UTC m=+156.137385776" watchObservedRunningTime="2025-11-24 14:21:59.026287107 +0000 UTC m=+156.142927584" Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.059080 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.061735 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.56169076 +0000 UTC m=+156.678331237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.066648 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" podStartSLOduration=132.066618148 podStartE2EDuration="2m12.066618148s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:59.063467187 +0000 UTC m=+156.180107694" watchObservedRunningTime="2025-11-24 14:21:59.066618148 +0000 UTC m=+156.183258625" Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.101914 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xtzhk" podStartSLOduration=133.101888228 podStartE2EDuration="2m13.101888228s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:59.101196805 +0000 UTC m=+156.217837302" watchObservedRunningTime="2025-11-24 14:21:59.101888228 +0000 UTC m=+156.218528705" Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.160825 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.161839 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.661818626 +0000 UTC m=+156.778459103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.263581 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.264058 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.76404296 +0000 UTC m=+156.880683437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.355441 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:21:59 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:21:59 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:21:59 crc kubenswrapper[4822]: healthz check failed Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.355510 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.366058 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.366436 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.866422368 +0000 UTC m=+156.983062845 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.467314 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.467836 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:21:59.967813865 +0000 UTC m=+157.084454342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.570823 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.571021 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.070988409 +0000 UTC m=+157.187628886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.571127 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.571594 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.071578508 +0000 UTC m=+157.188218985 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.671882 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.672433 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.172411917 +0000 UTC m=+157.289052384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.781970 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.782565 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.282538352 +0000 UTC m=+157.399178829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.888007 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.888809 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.388744713 +0000 UTC m=+157.505385190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.924756 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" event={"ID":"c9c4bffa-f289-4963-879e-1d54fc224dd3","Type":"ContainerStarted","Data":"cacb57effaac22b597e697eca96ff2e7c1b0c8ac2fdbc11b75c56fcdeb3f1b2e"} Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.924807 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" event={"ID":"c9c4bffa-f289-4963-879e-1d54fc224dd3","Type":"ContainerStarted","Data":"087bd08e18be62d2348c73550a37f54d0c2dbfba7b5fc2a777049e9e0afdab49"} Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.926035 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.935423 4822 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-dlp4b container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.935485 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" podUID="c9c4bffa-f289-4963-879e-1d54fc224dd3" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.941393 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" event={"ID":"28e02b8d-b939-4a4c-952e-3e8365a1d124","Type":"ContainerStarted","Data":"49db90a6345d7a7236495d88be66625d2956a3d3721b82656a7078cd2519b605"} Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.962835 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" podStartSLOduration=132.962813215 podStartE2EDuration="2m12.962813215s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:59.958647853 +0000 UTC m=+157.075288350" watchObservedRunningTime="2025-11-24 14:21:59.962813215 +0000 UTC m=+157.079453702" Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.969257 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r2lc6" event={"ID":"c019c6e3-3a0d-4838-a28d-2392d45c52b7","Type":"ContainerStarted","Data":"d260c05f1f97fee547c45ede184d6dab17f2ffaa37c60485be01621b8bbd4690"} Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.972848 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dq7pg" event={"ID":"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e","Type":"ContainerStarted","Data":"26063db8c42f6ac05f5f373e834463864c28d0798df9e5cb82083267db83fafc"} Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.985233 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" event={"ID":"70e35241-b7b3-4693-b7bb-a439ee947d03","Type":"ContainerStarted","Data":"49a30f4addd00b261191e947d6614cc1c5988544eda54bc88ba687e0d96361b3"} Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.990637 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:21:59 crc kubenswrapper[4822]: E1124 14:21:59.993737 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.493715975 +0000 UTC m=+157.610356452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:21:59 crc kubenswrapper[4822]: I1124 14:21:59.997876 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5w65w" podStartSLOduration=132.997846108 podStartE2EDuration="2m12.997846108s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:21:59.989380137 +0000 UTC m=+157.106020614" watchObservedRunningTime="2025-11-24 14:21:59.997846108 +0000 UTC m=+157.114486585" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.025333 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj" event={"ID":"0d499682-90a9-464e-9cb5-6c5a82155311","Type":"ContainerStarted","Data":"7eef8ca3cfb18708e359211bf17ab16b2a855b3aceb43a4aa1a371c2de103126"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.040881 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" event={"ID":"979147c0-9b48-4a0f-9506-00b31ca99603","Type":"ContainerStarted","Data":"3e0bf0a4ed5ea6616261881f3c0b078da04d750efe64fbda108e4847a7349ea5"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.052310 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" event={"ID":"3486a3b5-cba7-4d68-b377-310719458383","Type":"ContainerStarted","Data":"814ecfead93ec06014ea1c0e122fe271d904f8045ab0e4c2452b464ec3b40515"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.052372 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" event={"ID":"3486a3b5-cba7-4d68-b377-310719458383","Type":"ContainerStarted","Data":"0ed6b66ab17fd0ef0cfd317fc9b975df0d9a199faed87f52d0a43c58f3b19ddf"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.065809 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" event={"ID":"8d60f83e-1c5e-4839-89c4-fb49928e91a8","Type":"ContainerStarted","Data":"b436f884c9aa259616105fd7dad402a25f3e4ee62c906df2c4416b7482b6f170"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.066182 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmhjj" podStartSLOduration=133.066158155 podStartE2EDuration="2m13.066158155s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.064991948 +0000 UTC m=+157.181632425" watchObservedRunningTime="2025-11-24 14:22:00.066158155 +0000 UTC m=+157.182798632" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.068435 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b47sz" podStartSLOduration=133.068422947 podStartE2EDuration="2m13.068422947s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.030548425 +0000 UTC m=+157.147188902" watchObservedRunningTime="2025-11-24 14:22:00.068422947 +0000 UTC m=+157.185063434" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.086109 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" event={"ID":"773e3cb5-284d-448f-b025-65375eaa41c8","Type":"ContainerStarted","Data":"37e625a823a7ac1a62d545d582da6faf33c192480827cecc8024d94e22f8a812"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.086163 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" event={"ID":"773e3cb5-284d-448f-b025-65375eaa41c8","Type":"ContainerStarted","Data":"e6f270590dd102e350438a1a340c9cb154ed22d68468d23a93a9060cecd40b84"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.092048 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:00 crc kubenswrapper[4822]: E1124 14:22:00.093744 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.593722687 +0000 UTC m=+157.710363164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.108358 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" event={"ID":"81e0e571-ef24-47df-a054-54226fcfe214","Type":"ContainerStarted","Data":"daa98a83a2c2cb3913e8529636ba555e9e4ee1378bd6d90fc5b5985d331f6db7"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.108411 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" event={"ID":"81e0e571-ef24-47df-a054-54226fcfe214","Type":"ContainerStarted","Data":"d7e008b3750d1ccf0b92e9480b897f0a23a25d2fe13dfe717430c6194c96d04b"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.109478 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.110548 4822 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ftv2h container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.110588 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" podUID="81e0e571-ef24-47df-a054-54226fcfe214" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.115415 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" event={"ID":"8f7ff89a-1139-432a-9890-e1fe54077965","Type":"ContainerStarted","Data":"a1bd591f2f2b234bc8b5eb3f495d905604f16af7a8162c74938aeb241118ec3b"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.117571 4822 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hm5r8 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.117613 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" podUID="8f7ff89a-1139-432a-9890-e1fe54077965" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.151339 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" event={"ID":"2bf05124-0e4f-4262-95c6-d87ede36dcb5","Type":"ContainerStarted","Data":"92b46d93f67aaf5946ad748c16845702c0c8d7ef980781bec5ee908fa22b00d6"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.197278 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:00 crc kubenswrapper[4822]: E1124 14:22:00.197652 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.697636615 +0000 UTC m=+157.814277092 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.219789 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" event={"ID":"ae8e0e95-b76c-4353-b5af-2fd044fa9be3","Type":"ContainerStarted","Data":"cd9c650811bd197b60d46bb9b41af4ddfa69885019e4e4baba878fcd24d4ac4a"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.219842 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" event={"ID":"ae8e0e95-b76c-4353-b5af-2fd044fa9be3","Type":"ContainerStarted","Data":"dc20603516717c8ab145fb390f78a3da8eb586f466b36043503930576de831d7"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.237438 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" event={"ID":"15f981aa-3735-40f5-bc13-2cfad43ab9d8","Type":"ContainerStarted","Data":"b4c869d0bf0e5ad4e104c85ee6c7a01a76e21500090ce2ac23120e85d97e2435"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.255793 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qldnw" event={"ID":"f3212b9b-7c39-4243-b7c6-afd009c4f6df","Type":"ContainerStarted","Data":"5433ec606f532942500888e112809aaee672604971771d59cf1bf44e238a6008"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.255851 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qldnw" event={"ID":"f3212b9b-7c39-4243-b7c6-afd009c4f6df","Type":"ContainerStarted","Data":"1e7e772ab9e2a4c7d2d2c2875296a28175469059745e66e6c5cb1834f7242df9"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.257687 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hkt6b" podStartSLOduration=133.257662057 podStartE2EDuration="2m13.257662057s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.156509078 +0000 UTC m=+157.273149565" watchObservedRunningTime="2025-11-24 14:22:00.257662057 +0000 UTC m=+157.374302534" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.268510 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" event={"ID":"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5","Type":"ContainerStarted","Data":"229c03247751ca5f033bbedc32e030bcf331257b5de8920b6702feef257e616f"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.268564 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" event={"ID":"43181ae3-6a15-4ba4-b1e0-850ed07bb1c5","Type":"ContainerStarted","Data":"77aa65feeef8caf788b28e230e88ea8ed1d1d8e816752dafe21a0b3fcb2c0dff"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.297325 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" podStartSLOduration=134.297307657 podStartE2EDuration="2m14.297307657s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.258197375 +0000 UTC m=+157.374837862" watchObservedRunningTime="2025-11-24 14:22:00.297307657 +0000 UTC m=+157.413948134" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.299344 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:00 crc kubenswrapper[4822]: E1124 14:22:00.300717 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.800703425 +0000 UTC m=+157.917343892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.304632 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" event={"ID":"b934905c-f8df-4e00-9926-19df23d81e61","Type":"ContainerStarted","Data":"b53a62a1b79726b616e4d947f42760d84f585d31621811878f9bc8fbf633fe5d"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.313128 4822 generic.go:334] "Generic (PLEG): container finished" podID="e49fad7d-a39a-4270-9d25-50d381408838" containerID="3017bd10dde54eceb1395986447f74127ccbd0964116eb905e15b855358c890d" exitCode=0 Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.313260 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" event={"ID":"e49fad7d-a39a-4270-9d25-50d381408838","Type":"ContainerDied","Data":"3017bd10dde54eceb1395986447f74127ccbd0964116eb905e15b855358c890d"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.353360 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" event={"ID":"985f7b3a-f9d1-44cc-9132-6e24bc94e8f3","Type":"ContainerStarted","Data":"a80b52e5437c69ddc46edbafe5e13c18ee137cad5d02504ff01f85af24e7774a"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.353545 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vm8sf" podStartSLOduration=134.353508406 podStartE2EDuration="2m14.353508406s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.29866485 +0000 UTC m=+157.415305357" watchObservedRunningTime="2025-11-24 14:22:00.353508406 +0000 UTC m=+157.470148883" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.361124 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" podStartSLOduration=133.36110958 podStartE2EDuration="2m13.36110958s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.354535939 +0000 UTC m=+157.471176436" watchObservedRunningTime="2025-11-24 14:22:00.36110958 +0000 UTC m=+157.477750057" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.363824 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:00 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:00 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:00 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.363901 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.371805 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" event={"ID":"0442f63a-2e24-493c-9049-19bbfc0ee8ef","Type":"ContainerStarted","Data":"d7863bde690fd56bfc67c543d81e2411c52b8712783715bf70227d800ffbb444"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.385654 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" event={"ID":"f54974f4-d4d0-4962-baf5-d145bd8fc5f9","Type":"ContainerStarted","Data":"8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.385701 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.385712 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" event={"ID":"f54974f4-d4d0-4962-baf5-d145bd8fc5f9","Type":"ContainerStarted","Data":"316a77d30678ea484470c8402bf7857ba3b3d927a0600dbefbaf85f88f85d532"} Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.387489 4822 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-95knp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.388903 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" podUID="f54974f4-d4d0-4962-baf5-d145bd8fc5f9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.406416 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:00 crc kubenswrapper[4822]: E1124 14:22:00.411149 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:00.911124031 +0000 UTC m=+158.027764678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.427259 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" podStartSLOduration=133.427237117 podStartE2EDuration="2m13.427237117s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.425420809 +0000 UTC m=+157.542061306" watchObservedRunningTime="2025-11-24 14:22:00.427237117 +0000 UTC m=+157.543877594" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.481037 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7tt8s" podStartSLOduration=133.481020789 podStartE2EDuration="2m13.481020789s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.478464838 +0000 UTC m=+157.595105335" watchObservedRunningTime="2025-11-24 14:22:00.481020789 +0000 UTC m=+157.597661256" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.495657 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-k49s4" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.510979 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:00 crc kubenswrapper[4822]: E1124 14:22:00.511348 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.01132778 +0000 UTC m=+158.127968257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.543792 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qldnw" podStartSLOduration=7.543768269 podStartE2EDuration="7.543768269s" podCreationTimestamp="2025-11-24 14:21:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.540784053 +0000 UTC m=+157.657424540" watchObservedRunningTime="2025-11-24 14:22:00.543768269 +0000 UTC m=+157.660408746" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.605241 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z2jvf" podStartSLOduration=134.605222127 podStartE2EDuration="2m14.605222127s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.604953388 +0000 UTC m=+157.721593865" watchObservedRunningTime="2025-11-24 14:22:00.605222127 +0000 UTC m=+157.721862604" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.612773 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:00 crc kubenswrapper[4822]: E1124 14:22:00.613127 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.113115189 +0000 UTC m=+158.229755666 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.647294 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" podStartSLOduration=133.647269093 podStartE2EDuration="2m13.647269093s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.644658449 +0000 UTC m=+157.761298936" watchObservedRunningTime="2025-11-24 14:22:00.647269093 +0000 UTC m=+157.763909570" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.684026 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-v69sk" podStartSLOduration=133.683992278 podStartE2EDuration="2m13.683992278s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:00.680392714 +0000 UTC m=+157.797033191" watchObservedRunningTime="2025-11-24 14:22:00.683992278 +0000 UTC m=+157.800632745" Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.714982 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:00 crc kubenswrapper[4822]: E1124 14:22:00.715961 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.215938432 +0000 UTC m=+158.332579039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.822425 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:00 crc kubenswrapper[4822]: E1124 14:22:00.822858 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.322845585 +0000 UTC m=+158.439486062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:00 crc kubenswrapper[4822]: I1124 14:22:00.924651 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:00 crc kubenswrapper[4822]: E1124 14:22:00.925103 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.425044437 +0000 UTC m=+158.541684914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.026377 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.027339 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.527315863 +0000 UTC m=+158.643956340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.128141 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.128343 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.628316307 +0000 UTC m=+158.744956784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.128975 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.129351 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.629335959 +0000 UTC m=+158.745976436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.230197 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.230783 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.730759417 +0000 UTC m=+158.847399894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.233693 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.335418 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.335902 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.835881994 +0000 UTC m=+158.952522471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.350860 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:01 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:01 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:01 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.350949 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.391761 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" event={"ID":"15f981aa-3735-40f5-bc13-2cfad43ab9d8","Type":"ContainerStarted","Data":"47d53b9b2400e2b574dcac1ab7777ca3976cea494ec087b1a0ec2169ebc8c1de"} Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.394022 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dq7pg" event={"ID":"498b2ff4-3940-4b34-89d2-fa7d4dc96a5e","Type":"ContainerStarted","Data":"5438a4d5a2903ac80cbfd14f18cf6bba7dedffb397dc28f16705f2d8398b78d2"} Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.394196 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dq7pg" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.396528 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" event={"ID":"e49fad7d-a39a-4270-9d25-50d381408838","Type":"ContainerStarted","Data":"e322d5b34bbcfb843043b2854503333b48f9976866d7eac610621f228e3adc9d"} Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.398544 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-55vjd" event={"ID":"2bf05124-0e4f-4262-95c6-d87ede36dcb5","Type":"ContainerStarted","Data":"3e81491e228d8bc13806f11c7220901b681d71a0eaa361a83ec169486af761fb"} Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.400369 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" event={"ID":"8d60f83e-1c5e-4839-89c4-fb49928e91a8","Type":"ContainerStarted","Data":"d1fc8fc3eccb402edb0b6f81d90d6033be5f6787e2f6821a4b37723793bba79b"} Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.401238 4822 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ftv2h container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.401288 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" podUID="81e0e571-ef24-47df-a054-54226fcfe214" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.401384 4822 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-95knp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.401487 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" podUID="f54974f4-d4d0-4962-baf5-d145bd8fc5f9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.420470 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dlp4b" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.433107 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hm5r8" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.436903 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.437134 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.937096144 +0000 UTC m=+159.053736621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.437798 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.440903 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:01.940867185 +0000 UTC m=+159.057507822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.456874 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dq7pg" podStartSLOduration=9.456851797 podStartE2EDuration="9.456851797s" podCreationTimestamp="2025-11-24 14:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:01.456255578 +0000 UTC m=+158.572896085" watchObservedRunningTime="2025-11-24 14:22:01.456851797 +0000 UTC m=+158.573492274" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.459111 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" podStartSLOduration=135.459102189 podStartE2EDuration="2m15.459102189s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:01.426760103 +0000 UTC m=+158.543400610" watchObservedRunningTime="2025-11-24 14:22:01.459102189 +0000 UTC m=+158.575742666" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.540093 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.540613 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.040582498 +0000 UTC m=+159.157222975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.561872 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" podStartSLOduration=134.56184789899999 podStartE2EDuration="2m14.561847899s" podCreationTimestamp="2025-11-24 14:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:01.558900195 +0000 UTC m=+158.675540682" watchObservedRunningTime="2025-11-24 14:22:01.561847899 +0000 UTC m=+158.678488376" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.592596 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rnqdz" podStartSLOduration=135.592574033 podStartE2EDuration="2m15.592574033s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:01.586897631 +0000 UTC m=+158.703538108" watchObservedRunningTime="2025-11-24 14:22:01.592574033 +0000 UTC m=+158.709214510" Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.642373 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.642769 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.14275716 +0000 UTC m=+159.259397637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.743193 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.744822 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.244794898 +0000 UTC m=+159.361435375 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.846393 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.847072 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.347053412 +0000 UTC m=+159.463693889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:01 crc kubenswrapper[4822]: I1124 14:22:01.949003 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:01 crc kubenswrapper[4822]: E1124 14:22:01.949496 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.449474021 +0000 UTC m=+159.566114498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.051245 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.051750 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.551731006 +0000 UTC m=+159.668371483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.151996 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.152459 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.652440031 +0000 UTC m=+159.769080518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.253981 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.254358 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.754341044 +0000 UTC m=+159.870981521 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.262688 4822 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.351274 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:02 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:02 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:02 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.351404 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.354869 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.355260 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.855227244 +0000 UTC m=+159.971867721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.456485 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.456915 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:02.95690092 +0000 UTC m=+160.073541397 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.468635 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" event={"ID":"0442f63a-2e24-493c-9049-19bbfc0ee8ef","Type":"ContainerStarted","Data":"d87efa8974aa432aa03ea1b2866eb4b038e0174bd2b28775e3dfbea742e1e769"} Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.468681 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" event={"ID":"0442f63a-2e24-493c-9049-19bbfc0ee8ef","Type":"ContainerStarted","Data":"987ce2934e85c6a44bfa963d9173362734caf54cf0cc05ae7a6e17aec6bdaba4"} Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.480248 4822 generic.go:334] "Generic (PLEG): container finished" podID="979147c0-9b48-4a0f-9506-00b31ca99603" containerID="3e0bf0a4ed5ea6616261881f3c0b078da04d750efe64fbda108e4847a7349ea5" exitCode=0 Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.480561 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" event={"ID":"979147c0-9b48-4a0f-9506-00b31ca99603","Type":"ContainerDied","Data":"3e0bf0a4ed5ea6616261881f3c0b078da04d750efe64fbda108e4847a7349ea5"} Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.557838 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.558054 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:03.058015018 +0000 UTC m=+160.174655495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.558811 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.561475 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:03.061454438 +0000 UTC m=+160.178094915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.660063 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.660330 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:03.160296024 +0000 UTC m=+160.276936501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.660495 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.660832 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:03.1608205 +0000 UTC m=+160.277460977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.735411 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xjpf5"] Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.736721 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.739120 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.747395 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xjpf5"] Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.762040 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.762301 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:03.262241618 +0000 UTC m=+160.378882095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.762391 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-utilities\") pod \"certified-operators-xjpf5\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.762469 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-catalog-content\") pod \"certified-operators-xjpf5\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.762650 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.762885 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7p5k\" (UniqueName: \"kubernetes.io/projected/b002a922-0295-4465-a424-21c07d16a661-kube-api-access-j7p5k\") pod \"certified-operators-xjpf5\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.763120 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 14:22:03.263103146 +0000 UTC m=+160.379743623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2z5sq" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.833883 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.834707 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.836513 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.836842 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.844457 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.864138 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.864455 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.864499 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.864577 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-utilities\") pod \"certified-operators-xjpf5\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.864603 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-catalog-content\") pod \"certified-operators-xjpf5\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.864682 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7p5k\" (UniqueName: \"kubernetes.io/projected/b002a922-0295-4465-a424-21c07d16a661-kube-api-access-j7p5k\") pod \"certified-operators-xjpf5\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: E1124 14:22:02.865269 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 14:22:03.365249036 +0000 UTC m=+160.481889513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.865714 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-utilities\") pod \"certified-operators-xjpf5\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.865967 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-catalog-content\") pod \"certified-operators-xjpf5\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.887369 4822 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T14:22:02.263034292Z","Handler":null,"Name":""} Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.894043 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7p5k\" (UniqueName: \"kubernetes.io/projected/b002a922-0295-4465-a424-21c07d16a661-kube-api-access-j7p5k\") pod \"certified-operators-xjpf5\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.905717 4822 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.905768 4822 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.934129 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jhbbs"] Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.935435 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.937888 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.953549 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhbbs"] Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.966030 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.966100 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6zwv\" (UniqueName: \"kubernetes.io/projected/decf7f1e-5ed6-4026-ad3a-41144a216fd0-kube-api-access-n6zwv\") pod \"community-operators-jhbbs\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.966162 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-utilities\") pod \"community-operators-jhbbs\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.966239 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.966268 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.966351 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-catalog-content\") pod \"community-operators-jhbbs\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.967339 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.971970 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.972033 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:02 crc kubenswrapper[4822]: I1124 14:22:02.998937 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.035881 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2z5sq\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.053005 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.067151 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.067928 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-catalog-content\") pod \"community-operators-jhbbs\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.067962 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6zwv\" (UniqueName: \"kubernetes.io/projected/decf7f1e-5ed6-4026-ad3a-41144a216fd0-kube-api-access-n6zwv\") pod \"community-operators-jhbbs\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.067993 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-utilities\") pod \"community-operators-jhbbs\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.068568 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-utilities\") pod \"community-operators-jhbbs\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.069176 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-catalog-content\") pod \"community-operators-jhbbs\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.081862 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.089772 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6zwv\" (UniqueName: \"kubernetes.io/projected/decf7f1e-5ed6-4026-ad3a-41144a216fd0-kube-api-access-n6zwv\") pod \"community-operators-jhbbs\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.091320 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.132229 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zxg6h"] Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.133487 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.150632 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.155487 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zxg6h"] Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.168889 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-utilities\") pod \"certified-operators-zxg6h\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.168974 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6m7c\" (UniqueName: \"kubernetes.io/projected/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-kube-api-access-p6m7c\") pod \"certified-operators-zxg6h\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.169014 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-catalog-content\") pod \"certified-operators-zxg6h\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.257736 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.258671 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.270454 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-utilities\") pod \"certified-operators-zxg6h\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.270513 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6m7c\" (UniqueName: \"kubernetes.io/projected/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-kube-api-access-p6m7c\") pod \"certified-operators-zxg6h\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.270552 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-catalog-content\") pod \"certified-operators-zxg6h\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.271753 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-utilities\") pod \"certified-operators-zxg6h\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.274751 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-catalog-content\") pod \"certified-operators-zxg6h\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.305052 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6m7c\" (UniqueName: \"kubernetes.io/projected/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-kube-api-access-p6m7c\") pod \"certified-operators-zxg6h\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.323588 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pgcnl"] Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.324746 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.354945 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pgcnl"] Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.373351 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-catalog-content\") pod \"community-operators-pgcnl\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.373422 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-utilities\") pod \"community-operators-pgcnl\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.373473 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr2f6\" (UniqueName: \"kubernetes.io/projected/a80db4ed-d4de-49f6-95d7-513d109ba3e4-kube-api-access-jr2f6\") pod \"community-operators-pgcnl\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.389591 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:03 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:03 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:03 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.393244 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.433304 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xjpf5"] Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.453454 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.474733 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-catalog-content\") pod \"community-operators-pgcnl\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.474785 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-utilities\") pod \"community-operators-pgcnl\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.474816 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr2f6\" (UniqueName: \"kubernetes.io/projected/a80db4ed-d4de-49f6-95d7-513d109ba3e4-kube-api-access-jr2f6\") pod \"community-operators-pgcnl\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.475629 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-catalog-content\") pod \"community-operators-pgcnl\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.475837 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-utilities\") pod \"community-operators-pgcnl\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.538093 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr2f6\" (UniqueName: \"kubernetes.io/projected/a80db4ed-d4de-49f6-95d7-513d109ba3e4-kube-api-access-jr2f6\") pod \"community-operators-pgcnl\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.547765 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2z5sq"] Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.557075 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" event={"ID":"0442f63a-2e24-493c-9049-19bbfc0ee8ef","Type":"ContainerStarted","Data":"4f3470d7cc0596d85c168e2074457f56329741a5bcb0fb61649fe02f61c92598"} Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.573651 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjpf5" event={"ID":"b002a922-0295-4465-a424-21c07d16a661","Type":"ContainerStarted","Data":"2fc1dd559460aa3270b4f95a9a9c75daf740fe684768aff3ffbe5cd183aeb71d"} Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.585013 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-6rhjd" podStartSLOduration=11.584986543 podStartE2EDuration="11.584986543s" podCreationTimestamp="2025-11-24 14:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:03.580552492 +0000 UTC m=+160.697192969" watchObservedRunningTime="2025-11-24 14:22:03.584986543 +0000 UTC m=+160.701627030" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.687900 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhbbs"] Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.697084 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:03 crc kubenswrapper[4822]: W1124 14:22:03.717891 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddecf7f1e_5ed6_4026_ad3a_41144a216fd0.slice/crio-94b718c1a2d2e7fa3605d97e60de80b8e961c7237f5f274648a7623e819237e4 WatchSource:0}: Error finding container 94b718c1a2d2e7fa3605d97e60de80b8e961c7237f5f274648a7623e819237e4: Status 404 returned error can't find the container with id 94b718c1a2d2e7fa3605d97e60de80b8e961c7237f5f274648a7623e819237e4 Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.727692 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.728617 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 14:22:03 crc kubenswrapper[4822]: W1124 14:22:03.783502 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2ee77d04_4d9a_4d9d_a7fb_8f50c62bf3f9.slice/crio-bb07a899f54cd49d519c4f119c1500041115a45c12e68bcd6f167877a649c1ec WatchSource:0}: Error finding container bb07a899f54cd49d519c4f119c1500041115a45c12e68bcd6f167877a649c1ec: Status 404 returned error can't find the container with id bb07a899f54cd49d519c4f119c1500041115a45c12e68bcd6f167877a649c1ec Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.859170 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.868603 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zxg6h"] Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.891001 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/979147c0-9b48-4a0f-9506-00b31ca99603-config-volume\") pod \"979147c0-9b48-4a0f-9506-00b31ca99603\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.891142 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/979147c0-9b48-4a0f-9506-00b31ca99603-secret-volume\") pod \"979147c0-9b48-4a0f-9506-00b31ca99603\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.891166 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsk9q\" (UniqueName: \"kubernetes.io/projected/979147c0-9b48-4a0f-9506-00b31ca99603-kube-api-access-vsk9q\") pod \"979147c0-9b48-4a0f-9506-00b31ca99603\" (UID: \"979147c0-9b48-4a0f-9506-00b31ca99603\") " Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.892872 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/979147c0-9b48-4a0f-9506-00b31ca99603-config-volume" (OuterVolumeSpecName: "config-volume") pod "979147c0-9b48-4a0f-9506-00b31ca99603" (UID: "979147c0-9b48-4a0f-9506-00b31ca99603"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.898757 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/979147c0-9b48-4a0f-9506-00b31ca99603-kube-api-access-vsk9q" (OuterVolumeSpecName: "kube-api-access-vsk9q") pod "979147c0-9b48-4a0f-9506-00b31ca99603" (UID: "979147c0-9b48-4a0f-9506-00b31ca99603"). InnerVolumeSpecName "kube-api-access-vsk9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.899070 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/979147c0-9b48-4a0f-9506-00b31ca99603-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "979147c0-9b48-4a0f-9506-00b31ca99603" (UID: "979147c0-9b48-4a0f-9506-00b31ca99603"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.992434 4822 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/979147c0-9b48-4a0f-9506-00b31ca99603-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.992481 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsk9q\" (UniqueName: \"kubernetes.io/projected/979147c0-9b48-4a0f-9506-00b31ca99603-kube-api-access-vsk9q\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:03 crc kubenswrapper[4822]: I1124 14:22:03.992491 4822 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/979147c0-9b48-4a0f-9506-00b31ca99603-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.039429 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pgcnl"] Nov 24 14:22:04 crc kubenswrapper[4822]: W1124 14:22:04.049816 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda80db4ed_d4de_49f6_95d7_513d109ba3e4.slice/crio-cb25574e54a64a4e5e3bcd612d6e1c6b6575dded4807e3f4ef832f22aa141529 WatchSource:0}: Error finding container cb25574e54a64a4e5e3bcd612d6e1c6b6575dded4807e3f4ef832f22aa141529: Status 404 returned error can't find the container with id cb25574e54a64a4e5e3bcd612d6e1c6b6575dded4807e3f4ef832f22aa141529 Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.346575 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:04 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:04 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:04 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.346943 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.580109 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" event={"ID":"979147c0-9b48-4a0f-9506-00b31ca99603","Type":"ContainerDied","Data":"29618aa5e61a2deb1471f441c5cae592d8f4dbc7cf6158438651040c0b6d5951"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.580181 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29618aa5e61a2deb1471f441c5cae592d8f4dbc7cf6158438651040c0b6d5951" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.580141 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.582036 4822 generic.go:334] "Generic (PLEG): container finished" podID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerID="b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6" exitCode=0 Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.582122 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbbs" event={"ID":"decf7f1e-5ed6-4026-ad3a-41144a216fd0","Type":"ContainerDied","Data":"b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.582165 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbbs" event={"ID":"decf7f1e-5ed6-4026-ad3a-41144a216fd0","Type":"ContainerStarted","Data":"94b718c1a2d2e7fa3605d97e60de80b8e961c7237f5f274648a7623e819237e4"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.584690 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.585529 4822 generic.go:334] "Generic (PLEG): container finished" podID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerID="e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585" exitCode=0 Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.585567 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxg6h" event={"ID":"e000bd6c-9e10-4b1c-8989-9d08c8b7a705","Type":"ContainerDied","Data":"e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.585611 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxg6h" event={"ID":"e000bd6c-9e10-4b1c-8989-9d08c8b7a705","Type":"ContainerStarted","Data":"e98a4b58a4b98f71254c52f00109f30128e38472955d9d41ae9fe36e21f7df54"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.588074 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9","Type":"ContainerStarted","Data":"97834b25f320273ab84f24a6c191b48f5d97445cfb175f51249e3bdce44be6a7"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.588137 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9","Type":"ContainerStarted","Data":"bb07a899f54cd49d519c4f119c1500041115a45c12e68bcd6f167877a649c1ec"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.590439 4822 generic.go:334] "Generic (PLEG): container finished" podID="b002a922-0295-4465-a424-21c07d16a661" containerID="40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7" exitCode=0 Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.590516 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjpf5" event={"ID":"b002a922-0295-4465-a424-21c07d16a661","Type":"ContainerDied","Data":"40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.592471 4822 generic.go:334] "Generic (PLEG): container finished" podID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerID="dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096" exitCode=0 Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.592544 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pgcnl" event={"ID":"a80db4ed-d4de-49f6-95d7-513d109ba3e4","Type":"ContainerDied","Data":"dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.592573 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pgcnl" event={"ID":"a80db4ed-d4de-49f6-95d7-513d109ba3e4","Type":"ContainerStarted","Data":"cb25574e54a64a4e5e3bcd612d6e1c6b6575dded4807e3f4ef832f22aa141529"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.604747 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" event={"ID":"d5224ecc-afd0-4226-acd8-cb2c4197639c","Type":"ContainerStarted","Data":"b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.604811 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" event={"ID":"d5224ecc-afd0-4226-acd8-cb2c4197639c","Type":"ContainerStarted","Data":"3dab6393b3f64b698f899af77fc0e727f25b64ac63c9e0b5816dd18e3e8546fb"} Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.636958 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" podStartSLOduration=138.636937258 podStartE2EDuration="2m18.636937258s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:04.634126148 +0000 UTC m=+161.750766675" watchObservedRunningTime="2025-11-24 14:22:04.636937258 +0000 UTC m=+161.753577745" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.675453 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.675429691 podStartE2EDuration="2.675429691s" podCreationTimestamp="2025-11-24 14:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:04.656960839 +0000 UTC m=+161.773601316" watchObservedRunningTime="2025-11-24 14:22:04.675429691 +0000 UTC m=+161.792070178" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.922286 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lgxb4"] Nov 24 14:22:04 crc kubenswrapper[4822]: E1124 14:22:04.922618 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="979147c0-9b48-4a0f-9506-00b31ca99603" containerName="collect-profiles" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.922638 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="979147c0-9b48-4a0f-9506-00b31ca99603" containerName="collect-profiles" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.922798 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="979147c0-9b48-4a0f-9506-00b31ca99603" containerName="collect-profiles" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.923956 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.928607 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 14:22:04 crc kubenswrapper[4822]: I1124 14:22:04.928993 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lgxb4"] Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.008725 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-utilities\") pod \"redhat-marketplace-lgxb4\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.008794 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-catalog-content\") pod \"redhat-marketplace-lgxb4\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.009014 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b26fl\" (UniqueName: \"kubernetes.io/projected/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-kube-api-access-b26fl\") pod \"redhat-marketplace-lgxb4\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.075070 4822 patch_prober.go:28] interesting pod/downloads-7954f5f757-p9tw4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" start-of-body= Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.075138 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-p9tw4" podUID="911c455a-0bf1-4855-956f-34a7d3f2f080" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.075150 4822 patch_prober.go:28] interesting pod/downloads-7954f5f757-p9tw4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" start-of-body= Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.075195 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-p9tw4" podUID="911c455a-0bf1-4855-956f-34a7d3f2f080" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.110259 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b26fl\" (UniqueName: \"kubernetes.io/projected/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-kube-api-access-b26fl\") pod \"redhat-marketplace-lgxb4\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.110361 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-utilities\") pod \"redhat-marketplace-lgxb4\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.110385 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-catalog-content\") pod \"redhat-marketplace-lgxb4\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.110834 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-catalog-content\") pod \"redhat-marketplace-lgxb4\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.111063 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-utilities\") pod \"redhat-marketplace-lgxb4\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.136095 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b26fl\" (UniqueName: \"kubernetes.io/projected/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-kube-api-access-b26fl\") pod \"redhat-marketplace-lgxb4\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.235254 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.235655 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.237006 4822 patch_prober.go:28] interesting pod/console-f9d7485db-5v7wp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.237060 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5v7wp" podUID="6153af6d-c527-485f-8a0e-0cfecfb43d5b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.264701 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.325933 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5wc4g"] Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.329322 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.335584 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wc4g"] Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.350347 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.358414 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:05 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:05 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:05 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.358479 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.413959 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl7xs\" (UniqueName: \"kubernetes.io/projected/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-kube-api-access-vl7xs\") pod \"redhat-marketplace-5wc4g\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.414099 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-utilities\") pod \"redhat-marketplace-5wc4g\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.414132 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-catalog-content\") pod \"redhat-marketplace-5wc4g\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.496869 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lgxb4"] Nov 24 14:22:05 crc kubenswrapper[4822]: W1124 14:22:05.503006 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e5e4406_043a_46a3_8bf8_9de3cdbae55c.slice/crio-a8abda4e7f945c92a1ea3611a917efc2cb5064d4199660a12f60aaf846d35383 WatchSource:0}: Error finding container a8abda4e7f945c92a1ea3611a917efc2cb5064d4199660a12f60aaf846d35383: Status 404 returned error can't find the container with id a8abda4e7f945c92a1ea3611a917efc2cb5064d4199660a12f60aaf846d35383 Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.519020 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl7xs\" (UniqueName: \"kubernetes.io/projected/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-kube-api-access-vl7xs\") pod \"redhat-marketplace-5wc4g\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.519179 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-utilities\") pod \"redhat-marketplace-5wc4g\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.519219 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-catalog-content\") pod \"redhat-marketplace-5wc4g\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.519523 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.519718 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.519733 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-catalog-content\") pod \"redhat-marketplace-5wc4g\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.519962 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-utilities\") pod \"redhat-marketplace-5wc4g\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.530428 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.557840 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl7xs\" (UniqueName: \"kubernetes.io/projected/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-kube-api-access-vl7xs\") pod \"redhat-marketplace-5wc4g\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.614607 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lgxb4" event={"ID":"5e5e4406-043a-46a3-8bf8-9de3cdbae55c","Type":"ContainerStarted","Data":"a8abda4e7f945c92a1ea3611a917efc2cb5064d4199660a12f60aaf846d35383"} Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.617116 4822 generic.go:334] "Generic (PLEG): container finished" podID="2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9" containerID="97834b25f320273ab84f24a6c191b48f5d97445cfb175f51249e3bdce44be6a7" exitCode=0 Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.617954 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9","Type":"ContainerDied","Data":"97834b25f320273ab84f24a6c191b48f5d97445cfb175f51249e3bdce44be6a7"} Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.618109 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.634940 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8bc7n" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.655465 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.668459 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.847476 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.848156 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.874513 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.922618 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x55mp"] Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.923696 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:05 crc kubenswrapper[4822]: W1124 14:22:05.928890 4822 reflector.go:561] object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh": failed to list *v1.Secret: secrets "redhat-operators-dockercfg-ct8rh" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Nov 24 14:22:05 crc kubenswrapper[4822]: E1124 14:22:05.928962 4822 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-ct8rh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"redhat-operators-dockercfg-ct8rh\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 14:22:05 crc kubenswrapper[4822]: I1124 14:22:05.953780 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x55mp"] Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.038036 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-utilities\") pod \"redhat-operators-x55mp\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.038384 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxbtj\" (UniqueName: \"kubernetes.io/projected/b17ebf75-0076-48f5-898f-0196a270a555-kube-api-access-lxbtj\") pod \"redhat-operators-x55mp\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.038409 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-catalog-content\") pod \"redhat-operators-x55mp\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.140093 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxbtj\" (UniqueName: \"kubernetes.io/projected/b17ebf75-0076-48f5-898f-0196a270a555-kube-api-access-lxbtj\") pod \"redhat-operators-x55mp\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.140148 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-catalog-content\") pod \"redhat-operators-x55mp\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.140221 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-utilities\") pod \"redhat-operators-x55mp\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.141023 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-utilities\") pod \"redhat-operators-x55mp\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.141598 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-catalog-content\") pod \"redhat-operators-x55mp\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.176530 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxbtj\" (UniqueName: \"kubernetes.io/projected/b17ebf75-0076-48f5-898f-0196a270a555-kube-api-access-lxbtj\") pod \"redhat-operators-x55mp\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.218335 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wc4g"] Nov 24 14:22:06 crc kubenswrapper[4822]: W1124 14:22:06.289048 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod803cab3b_a4e9_4f87_a0a6_d3aa915c14de.slice/crio-cb088ca0d551ffa1ddec6e17829096e43383fecc0798afa126cee720aa6c74be WatchSource:0}: Error finding container cb088ca0d551ffa1ddec6e17829096e43383fecc0798afa126cee720aa6c74be: Status 404 returned error can't find the container with id cb088ca0d551ffa1ddec6e17829096e43383fecc0798afa126cee720aa6c74be Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.325018 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kkgmq"] Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.326320 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.338725 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kkgmq"] Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.347490 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:06 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:06 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:06 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.348071 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.352859 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.451127 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc624\" (UniqueName: \"kubernetes.io/projected/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-kube-api-access-rc624\") pod \"redhat-operators-kkgmq\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.451229 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-utilities\") pod \"redhat-operators-kkgmq\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.451287 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-catalog-content\") pod \"redhat-operators-kkgmq\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.555103 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc624\" (UniqueName: \"kubernetes.io/projected/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-kube-api-access-rc624\") pod \"redhat-operators-kkgmq\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.555985 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-utilities\") pod \"redhat-operators-kkgmq\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.556045 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-catalog-content\") pod \"redhat-operators-kkgmq\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.556843 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-utilities\") pod \"redhat-operators-kkgmq\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.557096 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-catalog-content\") pod \"redhat-operators-kkgmq\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.571345 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc624\" (UniqueName: \"kubernetes.io/projected/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-kube-api-access-rc624\") pod \"redhat-operators-kkgmq\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.624956 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wc4g" event={"ID":"803cab3b-a4e9-4f87-a0a6-d3aa915c14de","Type":"ContainerStarted","Data":"b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166"} Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.625300 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wc4g" event={"ID":"803cab3b-a4e9-4f87-a0a6-d3aa915c14de","Type":"ContainerStarted","Data":"cb088ca0d551ffa1ddec6e17829096e43383fecc0798afa126cee720aa6c74be"} Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.628080 4822 generic.go:334] "Generic (PLEG): container finished" podID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerID="263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd" exitCode=0 Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.629191 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lgxb4" event={"ID":"5e5e4406-043a-46a3-8bf8-9de3cdbae55c","Type":"ContainerDied","Data":"263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd"} Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.636042 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fdw9w" Nov 24 14:22:06 crc kubenswrapper[4822]: I1124 14:22:06.998271 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.168885 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kube-api-access\") pod \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\" (UID: \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\") " Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.168983 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kubelet-dir\") pod \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\" (UID: \"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9\") " Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.169465 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9" (UID: "2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.174036 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9" (UID: "2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.248878 4822 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/redhat-operators-x55mp" secret="" err="failed to sync secret cache: timed out waiting for the condition" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.249284 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.270069 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.270130 4822 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.320647 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.322015 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.357061 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 14:22:07 crc kubenswrapper[4822]: E1124 14:22:07.357520 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9" containerName="pruner" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.357604 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9" containerName="pruner" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.357762 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9" containerName="pruner" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.358254 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.361407 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.362767 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.362885 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:07 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:07 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:07 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.366737 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.362958 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.473545 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcca393b-facc-4010-95b7-9b1a0315a448-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"bcca393b-facc-4010-95b7-9b1a0315a448\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.473599 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcca393b-facc-4010-95b7-9b1a0315a448-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"bcca393b-facc-4010-95b7-9b1a0315a448\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.575030 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcca393b-facc-4010-95b7-9b1a0315a448-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"bcca393b-facc-4010-95b7-9b1a0315a448\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.575093 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcca393b-facc-4010-95b7-9b1a0315a448-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"bcca393b-facc-4010-95b7-9b1a0315a448\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.575252 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcca393b-facc-4010-95b7-9b1a0315a448-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"bcca393b-facc-4010-95b7-9b1a0315a448\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.610003 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcca393b-facc-4010-95b7-9b1a0315a448-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"bcca393b-facc-4010-95b7-9b1a0315a448\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.636752 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x55mp"] Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.645684 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2ee77d04-4d9a-4d9d-a7fb-8f50c62bf3f9","Type":"ContainerDied","Data":"bb07a899f54cd49d519c4f119c1500041115a45c12e68bcd6f167877a649c1ec"} Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.645729 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb07a899f54cd49d519c4f119c1500041115a45c12e68bcd6f167877a649c1ec" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.645698 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.655034 4822 generic.go:334] "Generic (PLEG): container finished" podID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerID="b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166" exitCode=0 Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.655315 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wc4g" event={"ID":"803cab3b-a4e9-4f87-a0a6-d3aa915c14de","Type":"ContainerDied","Data":"b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166"} Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.697479 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:07 crc kubenswrapper[4822]: I1124 14:22:07.725009 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kkgmq"] Nov 24 14:22:07 crc kubenswrapper[4822]: W1124 14:22:07.734825 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd14f6b2d_cb93_4fd0_acde_a1d42a0179a2.slice/crio-61a5fd793de702c19f2a880a3fe8038714808d228f66c31931e8c3370823b40d WatchSource:0}: Error finding container 61a5fd793de702c19f2a880a3fe8038714808d228f66c31931e8c3370823b40d: Status 404 returned error can't find the container with id 61a5fd793de702c19f2a880a3fe8038714808d228f66c31931e8c3370823b40d Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.039123 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 14:22:08 crc kubenswrapper[4822]: W1124 14:22:08.123744 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbcca393b_facc_4010_95b7_9b1a0315a448.slice/crio-bc767b899ee2867a346dcbd19b4388134be8dff1a2baee40eb2c7972422dc0ab WatchSource:0}: Error finding container bc767b899ee2867a346dcbd19b4388134be8dff1a2baee40eb2c7972422dc0ab: Status 404 returned error can't find the container with id bc767b899ee2867a346dcbd19b4388134be8dff1a2baee40eb2c7972422dc0ab Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.347568 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:08 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:08 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:08 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.347659 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.688878 4822 generic.go:334] "Generic (PLEG): container finished" podID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerID="a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032" exitCode=0 Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.689221 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkgmq" event={"ID":"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2","Type":"ContainerDied","Data":"a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032"} Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.689256 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkgmq" event={"ID":"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2","Type":"ContainerStarted","Data":"61a5fd793de702c19f2a880a3fe8038714808d228f66c31931e8c3370823b40d"} Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.707915 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bcca393b-facc-4010-95b7-9b1a0315a448","Type":"ContainerStarted","Data":"b77b3a764be32a7734ff1c856a862fd694e0d46c49f1e27196dec729dc57340c"} Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.707974 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bcca393b-facc-4010-95b7-9b1a0315a448","Type":"ContainerStarted","Data":"bc767b899ee2867a346dcbd19b4388134be8dff1a2baee40eb2c7972422dc0ab"} Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.735258 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=1.735238533 podStartE2EDuration="1.735238533s" podCreationTimestamp="2025-11-24 14:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:08.734633414 +0000 UTC m=+165.851273901" watchObservedRunningTime="2025-11-24 14:22:08.735238533 +0000 UTC m=+165.851879010" Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.738786 4822 generic.go:334] "Generic (PLEG): container finished" podID="b17ebf75-0076-48f5-898f-0196a270a555" containerID="cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e" exitCode=0 Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.739003 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55mp" event={"ID":"b17ebf75-0076-48f5-898f-0196a270a555","Type":"ContainerDied","Data":"cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e"} Nov 24 14:22:08 crc kubenswrapper[4822]: I1124 14:22:08.739131 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55mp" event={"ID":"b17ebf75-0076-48f5-898f-0196a270a555","Type":"ContainerStarted","Data":"acd3be21cebe24df5ef922fa97f3f6bf5b765628babf94ddc1c355459393d256"} Nov 24 14:22:09 crc kubenswrapper[4822]: I1124 14:22:09.351586 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:09 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:09 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:09 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:09 crc kubenswrapper[4822]: I1124 14:22:09.351656 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:09 crc kubenswrapper[4822]: I1124 14:22:09.619035 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:22:09 crc kubenswrapper[4822]: I1124 14:22:09.641348 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b8d0fb3c-6504-4242-b078-ee8c88e7f6f7-metrics-certs\") pod \"network-metrics-daemon-dljmt\" (UID: \"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7\") " pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:22:09 crc kubenswrapper[4822]: I1124 14:22:09.743109 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dljmt" Nov 24 14:22:09 crc kubenswrapper[4822]: I1124 14:22:09.785810 4822 generic.go:334] "Generic (PLEG): container finished" podID="bcca393b-facc-4010-95b7-9b1a0315a448" containerID="b77b3a764be32a7734ff1c856a862fd694e0d46c49f1e27196dec729dc57340c" exitCode=0 Nov 24 14:22:09 crc kubenswrapper[4822]: I1124 14:22:09.785856 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bcca393b-facc-4010-95b7-9b1a0315a448","Type":"ContainerDied","Data":"b77b3a764be32a7734ff1c856a862fd694e0d46c49f1e27196dec729dc57340c"} Nov 24 14:22:10 crc kubenswrapper[4822]: I1124 14:22:10.357552 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:10 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:10 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:10 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:10 crc kubenswrapper[4822]: I1124 14:22:10.357951 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:10 crc kubenswrapper[4822]: I1124 14:22:10.479961 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-dljmt"] Nov 24 14:22:10 crc kubenswrapper[4822]: I1124 14:22:10.811628 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-dljmt" event={"ID":"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7","Type":"ContainerStarted","Data":"e0774b27b32b6f49d60be1750d9bc2ae3c9efa2ebb220ae97f7e5b153cd59df2"} Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.126540 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dq7pg" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.127726 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.254016 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcca393b-facc-4010-95b7-9b1a0315a448-kube-api-access\") pod \"bcca393b-facc-4010-95b7-9b1a0315a448\" (UID: \"bcca393b-facc-4010-95b7-9b1a0315a448\") " Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.254464 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcca393b-facc-4010-95b7-9b1a0315a448-kubelet-dir\") pod \"bcca393b-facc-4010-95b7-9b1a0315a448\" (UID: \"bcca393b-facc-4010-95b7-9b1a0315a448\") " Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.254626 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bcca393b-facc-4010-95b7-9b1a0315a448-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bcca393b-facc-4010-95b7-9b1a0315a448" (UID: "bcca393b-facc-4010-95b7-9b1a0315a448"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.255364 4822 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcca393b-facc-4010-95b7-9b1a0315a448-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.259926 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcca393b-facc-4010-95b7-9b1a0315a448-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bcca393b-facc-4010-95b7-9b1a0315a448" (UID: "bcca393b-facc-4010-95b7-9b1a0315a448"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.347394 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:11 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:11 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:11 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.347452 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.357088 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcca393b-facc-4010-95b7-9b1a0315a448-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.578329 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.578389 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.823522 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.823517 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bcca393b-facc-4010-95b7-9b1a0315a448","Type":"ContainerDied","Data":"bc767b899ee2867a346dcbd19b4388134be8dff1a2baee40eb2c7972422dc0ab"} Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.823579 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc767b899ee2867a346dcbd19b4388134be8dff1a2baee40eb2c7972422dc0ab" Nov 24 14:22:11 crc kubenswrapper[4822]: I1124 14:22:11.827763 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-dljmt" event={"ID":"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7","Type":"ContainerStarted","Data":"6fc0bb36f46cd240c124bf5f681992718b0ddc6b730c4d5bb0c615b663f2e87b"} Nov 24 14:22:12 crc kubenswrapper[4822]: I1124 14:22:12.347008 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:12 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:12 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:12 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:12 crc kubenswrapper[4822]: I1124 14:22:12.347089 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:13 crc kubenswrapper[4822]: I1124 14:22:13.346917 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:13 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:13 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:13 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:13 crc kubenswrapper[4822]: I1124 14:22:13.347326 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:14 crc kubenswrapper[4822]: I1124 14:22:14.346645 4822 patch_prober.go:28] interesting pod/router-default-5444994796-w4gtt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 14:22:14 crc kubenswrapper[4822]: [-]has-synced failed: reason withheld Nov 24 14:22:14 crc kubenswrapper[4822]: [+]process-running ok Nov 24 14:22:14 crc kubenswrapper[4822]: healthz check failed Nov 24 14:22:14 crc kubenswrapper[4822]: I1124 14:22:14.346726 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4gtt" podUID="338efc62-c0d6-4589-a6a8-dd783e0f08ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 14:22:15 crc kubenswrapper[4822]: I1124 14:22:15.080010 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-p9tw4" Nov 24 14:22:15 crc kubenswrapper[4822]: I1124 14:22:15.235770 4822 patch_prober.go:28] interesting pod/console-f9d7485db-5v7wp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 24 14:22:15 crc kubenswrapper[4822]: I1124 14:22:15.235837 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5v7wp" podUID="6153af6d-c527-485f-8a0e-0cfecfb43d5b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 24 14:22:15 crc kubenswrapper[4822]: I1124 14:22:15.347617 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:22:15 crc kubenswrapper[4822]: I1124 14:22:15.349991 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-w4gtt" Nov 24 14:22:23 crc kubenswrapper[4822]: I1124 14:22:23.097625 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:22:25 crc kubenswrapper[4822]: I1124 14:22:25.239658 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:22:25 crc kubenswrapper[4822]: I1124 14:22:25.243405 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.947706 4822 generic.go:334] "Generic (PLEG): container finished" podID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerID="10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab" exitCode=0 Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.947812 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbbs" event={"ID":"decf7f1e-5ed6-4026-ad3a-41144a216fd0","Type":"ContainerDied","Data":"10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab"} Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.951591 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55mp" event={"ID":"b17ebf75-0076-48f5-898f-0196a270a555","Type":"ContainerStarted","Data":"1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8"} Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.963379 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-dljmt" event={"ID":"b8d0fb3c-6504-4242-b078-ee8c88e7f6f7","Type":"ContainerStarted","Data":"eb1f6d1cf4e163e297bc181a152791c27211ed98d4572d1a9fac33c0ca031b61"} Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.972519 4822 generic.go:334] "Generic (PLEG): container finished" podID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerID="d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40" exitCode=0 Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.972608 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lgxb4" event={"ID":"5e5e4406-043a-46a3-8bf8-9de3cdbae55c","Type":"ContainerDied","Data":"d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40"} Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.979154 4822 generic.go:334] "Generic (PLEG): container finished" podID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerID="9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e" exitCode=0 Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.979246 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkgmq" event={"ID":"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2","Type":"ContainerDied","Data":"9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e"} Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.982750 4822 generic.go:334] "Generic (PLEG): container finished" podID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerID="8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c" exitCode=0 Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.982806 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pgcnl" event={"ID":"a80db4ed-d4de-49f6-95d7-513d109ba3e4","Type":"ContainerDied","Data":"8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c"} Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.989167 4822 generic.go:334] "Generic (PLEG): container finished" podID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerID="260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875" exitCode=0 Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.989307 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxg6h" event={"ID":"e000bd6c-9e10-4b1c-8989-9d08c8b7a705","Type":"ContainerDied","Data":"260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875"} Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.994767 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-dljmt" podStartSLOduration=164.994750208 podStartE2EDuration="2m44.994750208s" podCreationTimestamp="2025-11-24 14:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:22:30.993650362 +0000 UTC m=+188.110290869" watchObservedRunningTime="2025-11-24 14:22:30.994750208 +0000 UTC m=+188.111390685" Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.999198 4822 generic.go:334] "Generic (PLEG): container finished" podID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerID="25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2" exitCode=0 Nov 24 14:22:30 crc kubenswrapper[4822]: I1124 14:22:30.999313 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wc4g" event={"ID":"803cab3b-a4e9-4f87-a0a6-d3aa915c14de","Type":"ContainerDied","Data":"25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2"} Nov 24 14:22:31 crc kubenswrapper[4822]: I1124 14:22:31.003990 4822 generic.go:334] "Generic (PLEG): container finished" podID="b002a922-0295-4465-a424-21c07d16a661" containerID="25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5" exitCode=0 Nov 24 14:22:31 crc kubenswrapper[4822]: I1124 14:22:31.004022 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjpf5" event={"ID":"b002a922-0295-4465-a424-21c07d16a661","Type":"ContainerDied","Data":"25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5"} Nov 24 14:22:31 crc kubenswrapper[4822]: I1124 14:22:31.758998 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 14:22:32 crc kubenswrapper[4822]: I1124 14:22:32.021484 4822 generic.go:334] "Generic (PLEG): container finished" podID="b17ebf75-0076-48f5-898f-0196a270a555" containerID="1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8" exitCode=0 Nov 24 14:22:32 crc kubenswrapper[4822]: I1124 14:22:32.021634 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55mp" event={"ID":"b17ebf75-0076-48f5-898f-0196a270a555","Type":"ContainerDied","Data":"1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8"} Nov 24 14:22:32 crc kubenswrapper[4822]: I1124 14:22:32.027078 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjpf5" event={"ID":"b002a922-0295-4465-a424-21c07d16a661","Type":"ContainerStarted","Data":"67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d"} Nov 24 14:22:32 crc kubenswrapper[4822]: I1124 14:22:32.029680 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lgxb4" event={"ID":"5e5e4406-043a-46a3-8bf8-9de3cdbae55c","Type":"ContainerStarted","Data":"2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934"} Nov 24 14:22:32 crc kubenswrapper[4822]: I1124 14:22:32.031925 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkgmq" event={"ID":"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2","Type":"ContainerStarted","Data":"2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c"} Nov 24 14:22:32 crc kubenswrapper[4822]: I1124 14:22:32.084025 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xjpf5" podStartSLOduration=2.991599045 podStartE2EDuration="30.084004246s" podCreationTimestamp="2025-11-24 14:22:02 +0000 UTC" firstStartedPulling="2025-11-24 14:22:04.593498268 +0000 UTC m=+161.710138775" lastFinishedPulling="2025-11-24 14:22:31.685903489 +0000 UTC m=+188.802543976" observedRunningTime="2025-11-24 14:22:32.067952643 +0000 UTC m=+189.184593120" watchObservedRunningTime="2025-11-24 14:22:32.084004246 +0000 UTC m=+189.200644713" Nov 24 14:22:32 crc kubenswrapper[4822]: I1124 14:22:32.086633 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lgxb4" podStartSLOduration=2.978229681 podStartE2EDuration="28.08662189s" podCreationTimestamp="2025-11-24 14:22:04 +0000 UTC" firstStartedPulling="2025-11-24 14:22:06.630074163 +0000 UTC m=+163.746714640" lastFinishedPulling="2025-11-24 14:22:31.738466372 +0000 UTC m=+188.855106849" observedRunningTime="2025-11-24 14:22:32.082892271 +0000 UTC m=+189.199532748" watchObservedRunningTime="2025-11-24 14:22:32.08662189 +0000 UTC m=+189.203262367" Nov 24 14:22:32 crc kubenswrapper[4822]: I1124 14:22:32.101765 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kkgmq" podStartSLOduration=3.092014518 podStartE2EDuration="26.101740565s" podCreationTimestamp="2025-11-24 14:22:06 +0000 UTC" firstStartedPulling="2025-11-24 14:22:08.699058954 +0000 UTC m=+165.815699431" lastFinishedPulling="2025-11-24 14:22:31.708785001 +0000 UTC m=+188.825425478" observedRunningTime="2025-11-24 14:22:32.099892676 +0000 UTC m=+189.216533173" watchObservedRunningTime="2025-11-24 14:22:32.101740565 +0000 UTC m=+189.218381062" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.039953 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbbs" event={"ID":"decf7f1e-5ed6-4026-ad3a-41144a216fd0","Type":"ContainerStarted","Data":"9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2"} Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.043903 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55mp" event={"ID":"b17ebf75-0076-48f5-898f-0196a270a555","Type":"ContainerStarted","Data":"99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf"} Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.046074 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxg6h" event={"ID":"e000bd6c-9e10-4b1c-8989-9d08c8b7a705","Type":"ContainerStarted","Data":"b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452"} Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.048443 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wc4g" event={"ID":"803cab3b-a4e9-4f87-a0a6-d3aa915c14de","Type":"ContainerStarted","Data":"e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6"} Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.051017 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pgcnl" event={"ID":"a80db4ed-d4de-49f6-95d7-513d109ba3e4","Type":"ContainerStarted","Data":"0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637"} Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.062892 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.063152 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.077938 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jhbbs" podStartSLOduration=3.592163846 podStartE2EDuration="31.077891302s" podCreationTimestamp="2025-11-24 14:22:02 +0000 UTC" firstStartedPulling="2025-11-24 14:22:04.584259812 +0000 UTC m=+161.700900309" lastFinishedPulling="2025-11-24 14:22:32.069987288 +0000 UTC m=+189.186627765" observedRunningTime="2025-11-24 14:22:33.076000702 +0000 UTC m=+190.192641199" watchObservedRunningTime="2025-11-24 14:22:33.077891302 +0000 UTC m=+190.194531789" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.096729 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zxg6h" podStartSLOduration=2.535135681 podStartE2EDuration="30.096712566s" podCreationTimestamp="2025-11-24 14:22:03 +0000 UTC" firstStartedPulling="2025-11-24 14:22:04.587514776 +0000 UTC m=+161.704155263" lastFinishedPulling="2025-11-24 14:22:32.149091671 +0000 UTC m=+189.265732148" observedRunningTime="2025-11-24 14:22:33.096581462 +0000 UTC m=+190.213221959" watchObservedRunningTime="2025-11-24 14:22:33.096712566 +0000 UTC m=+190.213353043" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.117527 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5wc4g" podStartSLOduration=3.6373287469999998 podStartE2EDuration="28.117509371s" podCreationTimestamp="2025-11-24 14:22:05 +0000 UTC" firstStartedPulling="2025-11-24 14:22:07.658220095 +0000 UTC m=+164.774860572" lastFinishedPulling="2025-11-24 14:22:32.138400719 +0000 UTC m=+189.255041196" observedRunningTime="2025-11-24 14:22:33.114906888 +0000 UTC m=+190.231547365" watchObservedRunningTime="2025-11-24 14:22:33.117509371 +0000 UTC m=+190.234149848" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.134835 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pgcnl" podStartSLOduration=2.728884486 podStartE2EDuration="30.134817166s" podCreationTimestamp="2025-11-24 14:22:03 +0000 UTC" firstStartedPulling="2025-11-24 14:22:04.594003654 +0000 UTC m=+161.710644131" lastFinishedPulling="2025-11-24 14:22:31.999936334 +0000 UTC m=+189.116576811" observedRunningTime="2025-11-24 14:22:33.133762001 +0000 UTC m=+190.250402508" watchObservedRunningTime="2025-11-24 14:22:33.134817166 +0000 UTC m=+190.251457643" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.152569 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x55mp" podStartSLOduration=4.329391999 podStartE2EDuration="28.152548234s" podCreationTimestamp="2025-11-24 14:22:05 +0000 UTC" firstStartedPulling="2025-11-24 14:22:08.742373961 +0000 UTC m=+165.859014438" lastFinishedPulling="2025-11-24 14:22:32.565530196 +0000 UTC m=+189.682170673" observedRunningTime="2025-11-24 14:22:33.15118538 +0000 UTC m=+190.267825867" watchObservedRunningTime="2025-11-24 14:22:33.152548234 +0000 UTC m=+190.269188711" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.258319 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.258375 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.454845 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.454912 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.698186 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:33 crc kubenswrapper[4822]: I1124 14:22:33.698273 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:34 crc kubenswrapper[4822]: I1124 14:22:34.212826 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-xjpf5" podUID="b002a922-0295-4465-a424-21c07d16a661" containerName="registry-server" probeResult="failure" output=< Nov 24 14:22:34 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:22:34 crc kubenswrapper[4822]: > Nov 24 14:22:34 crc kubenswrapper[4822]: I1124 14:22:34.292781 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jhbbs" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerName="registry-server" probeResult="failure" output=< Nov 24 14:22:34 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:22:34 crc kubenswrapper[4822]: > Nov 24 14:22:34 crc kubenswrapper[4822]: I1124 14:22:34.519267 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-zxg6h" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerName="registry-server" probeResult="failure" output=< Nov 24 14:22:34 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:22:34 crc kubenswrapper[4822]: > Nov 24 14:22:34 crc kubenswrapper[4822]: I1124 14:22:34.733634 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-pgcnl" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerName="registry-server" probeResult="failure" output=< Nov 24 14:22:34 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:22:34 crc kubenswrapper[4822]: > Nov 24 14:22:34 crc kubenswrapper[4822]: I1124 14:22:34.744058 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gzxq"] Nov 24 14:22:35 crc kubenswrapper[4822]: I1124 14:22:35.265054 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:35 crc kubenswrapper[4822]: I1124 14:22:35.265147 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:35 crc kubenswrapper[4822]: I1124 14:22:35.320391 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:35 crc kubenswrapper[4822]: I1124 14:22:35.656141 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:35 crc kubenswrapper[4822]: I1124 14:22:35.657353 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:35 crc kubenswrapper[4822]: I1124 14:22:35.697023 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:35 crc kubenswrapper[4822]: I1124 14:22:35.729444 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ksvsx" Nov 24 14:22:36 crc kubenswrapper[4822]: I1124 14:22:36.102978 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:22:37 crc kubenswrapper[4822]: I1124 14:22:37.251453 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:37 crc kubenswrapper[4822]: I1124 14:22:37.251527 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:37 crc kubenswrapper[4822]: I1124 14:22:37.323657 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:37 crc kubenswrapper[4822]: I1124 14:22:37.323835 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:38 crc kubenswrapper[4822]: I1124 14:22:38.290340 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x55mp" podUID="b17ebf75-0076-48f5-898f-0196a270a555" containerName="registry-server" probeResult="failure" output=< Nov 24 14:22:38 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:22:38 crc kubenswrapper[4822]: > Nov 24 14:22:38 crc kubenswrapper[4822]: I1124 14:22:38.369681 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kkgmq" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerName="registry-server" probeResult="failure" output=< Nov 24 14:22:38 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:22:38 crc kubenswrapper[4822]: > Nov 24 14:22:41 crc kubenswrapper[4822]: I1124 14:22:41.577878 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:22:41 crc kubenswrapper[4822]: I1124 14:22:41.578166 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:22:43 crc kubenswrapper[4822]: I1124 14:22:43.114162 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:43 crc kubenswrapper[4822]: I1124 14:22:43.181560 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:22:43 crc kubenswrapper[4822]: I1124 14:22:43.295901 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:43 crc kubenswrapper[4822]: I1124 14:22:43.337895 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:22:43 crc kubenswrapper[4822]: I1124 14:22:43.496018 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:43 crc kubenswrapper[4822]: I1124 14:22:43.537219 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:43 crc kubenswrapper[4822]: I1124 14:22:43.748793 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:43 crc kubenswrapper[4822]: I1124 14:22:43.788577 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.163923 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pgcnl"] Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.165038 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pgcnl" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerName="registry-server" containerID="cri-o://0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637" gracePeriod=2 Nov 24 14:22:45 crc kubenswrapper[4822]: E1124 14:22:45.342934 4822 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda80db4ed_d4de_49f6_95d7_513d109ba3e4.slice/crio-conmon-0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637.scope\": RecentStats: unable to find data in memory cache]" Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.618274 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.713497 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.763032 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zxg6h"] Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.763593 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zxg6h" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerName="registry-server" containerID="cri-o://b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452" gracePeriod=2 Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.801643 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr2f6\" (UniqueName: \"kubernetes.io/projected/a80db4ed-d4de-49f6-95d7-513d109ba3e4-kube-api-access-jr2f6\") pod \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.801738 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-utilities\") pod \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.801780 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-catalog-content\") pod \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\" (UID: \"a80db4ed-d4de-49f6-95d7-513d109ba3e4\") " Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.803105 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-utilities" (OuterVolumeSpecName: "utilities") pod "a80db4ed-d4de-49f6-95d7-513d109ba3e4" (UID: "a80db4ed-d4de-49f6-95d7-513d109ba3e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.810764 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a80db4ed-d4de-49f6-95d7-513d109ba3e4-kube-api-access-jr2f6" (OuterVolumeSpecName: "kube-api-access-jr2f6") pod "a80db4ed-d4de-49f6-95d7-513d109ba3e4" (UID: "a80db4ed-d4de-49f6-95d7-513d109ba3e4"). InnerVolumeSpecName "kube-api-access-jr2f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.853145 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a80db4ed-d4de-49f6-95d7-513d109ba3e4" (UID: "a80db4ed-d4de-49f6-95d7-513d109ba3e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.903573 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.903621 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80db4ed-d4de-49f6-95d7-513d109ba3e4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:45 crc kubenswrapper[4822]: I1124 14:22:45.903637 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr2f6\" (UniqueName: \"kubernetes.io/projected/a80db4ed-d4de-49f6-95d7-513d109ba3e4-kube-api-access-jr2f6\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.076805 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.127265 4822 generic.go:334] "Generic (PLEG): container finished" podID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerID="b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452" exitCode=0 Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.127323 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxg6h" event={"ID":"e000bd6c-9e10-4b1c-8989-9d08c8b7a705","Type":"ContainerDied","Data":"b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452"} Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.127365 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zxg6h" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.127396 4822 scope.go:117] "RemoveContainer" containerID="b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.127379 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxg6h" event={"ID":"e000bd6c-9e10-4b1c-8989-9d08c8b7a705","Type":"ContainerDied","Data":"e98a4b58a4b98f71254c52f00109f30128e38472955d9d41ae9fe36e21f7df54"} Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.134826 4822 generic.go:334] "Generic (PLEG): container finished" podID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerID="0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637" exitCode=0 Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.134871 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pgcnl" event={"ID":"a80db4ed-d4de-49f6-95d7-513d109ba3e4","Type":"ContainerDied","Data":"0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637"} Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.134903 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pgcnl" event={"ID":"a80db4ed-d4de-49f6-95d7-513d109ba3e4","Type":"ContainerDied","Data":"cb25574e54a64a4e5e3bcd612d6e1c6b6575dded4807e3f4ef832f22aa141529"} Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.134954 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pgcnl" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.144775 4822 scope.go:117] "RemoveContainer" containerID="260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.167532 4822 scope.go:117] "RemoveContainer" containerID="e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.176501 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pgcnl"] Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.183394 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pgcnl"] Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.197732 4822 scope.go:117] "RemoveContainer" containerID="b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452" Nov 24 14:22:46 crc kubenswrapper[4822]: E1124 14:22:46.198381 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452\": container with ID starting with b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452 not found: ID does not exist" containerID="b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.198437 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452"} err="failed to get container status \"b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452\": rpc error: code = NotFound desc = could not find container \"b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452\": container with ID starting with b7eace5d9bac715fa98c2f74ce606788d07843b15ba74bbcf7c1f6687282a452 not found: ID does not exist" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.198506 4822 scope.go:117] "RemoveContainer" containerID="260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875" Nov 24 14:22:46 crc kubenswrapper[4822]: E1124 14:22:46.199061 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875\": container with ID starting with 260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875 not found: ID does not exist" containerID="260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.199091 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875"} err="failed to get container status \"260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875\": rpc error: code = NotFound desc = could not find container \"260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875\": container with ID starting with 260cbb37a298a500ef59ec131853413acbba1ba7818a6054f553adb480b6d875 not found: ID does not exist" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.199116 4822 scope.go:117] "RemoveContainer" containerID="e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585" Nov 24 14:22:46 crc kubenswrapper[4822]: E1124 14:22:46.199343 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585\": container with ID starting with e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585 not found: ID does not exist" containerID="e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.199375 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585"} err="failed to get container status \"e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585\": rpc error: code = NotFound desc = could not find container \"e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585\": container with ID starting with e86b9e32aa7b75d99a563547d5e049e1c28097d8849094b12e53061611fe8585 not found: ID does not exist" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.199391 4822 scope.go:117] "RemoveContainer" containerID="0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.206882 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-catalog-content\") pod \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.207002 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-utilities\") pod \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.207036 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6m7c\" (UniqueName: \"kubernetes.io/projected/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-kube-api-access-p6m7c\") pod \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\" (UID: \"e000bd6c-9e10-4b1c-8989-9d08c8b7a705\") " Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.208540 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-utilities" (OuterVolumeSpecName: "utilities") pod "e000bd6c-9e10-4b1c-8989-9d08c8b7a705" (UID: "e000bd6c-9e10-4b1c-8989-9d08c8b7a705"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.211612 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-kube-api-access-p6m7c" (OuterVolumeSpecName: "kube-api-access-p6m7c") pod "e000bd6c-9e10-4b1c-8989-9d08c8b7a705" (UID: "e000bd6c-9e10-4b1c-8989-9d08c8b7a705"). InnerVolumeSpecName "kube-api-access-p6m7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.212051 4822 scope.go:117] "RemoveContainer" containerID="8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.229563 4822 scope.go:117] "RemoveContainer" containerID="dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.253073 4822 scope.go:117] "RemoveContainer" containerID="0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637" Nov 24 14:22:46 crc kubenswrapper[4822]: E1124 14:22:46.253711 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637\": container with ID starting with 0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637 not found: ID does not exist" containerID="0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.253771 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637"} err="failed to get container status \"0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637\": rpc error: code = NotFound desc = could not find container \"0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637\": container with ID starting with 0bdec25998e8b161d604f900696dda682326bffbafb8989a440e9941f0718637 not found: ID does not exist" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.253806 4822 scope.go:117] "RemoveContainer" containerID="8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c" Nov 24 14:22:46 crc kubenswrapper[4822]: E1124 14:22:46.254265 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c\": container with ID starting with 8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c not found: ID does not exist" containerID="8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.254347 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c"} err="failed to get container status \"8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c\": rpc error: code = NotFound desc = could not find container \"8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c\": container with ID starting with 8a85c7cf82ce0f37b2c2810e74b2fa1202f90c01dd1768d21ff7d68ad166fc0c not found: ID does not exist" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.254371 4822 scope.go:117] "RemoveContainer" containerID="dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096" Nov 24 14:22:46 crc kubenswrapper[4822]: E1124 14:22:46.255234 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096\": container with ID starting with dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096 not found: ID does not exist" containerID="dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.255286 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096"} err="failed to get container status \"dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096\": rpc error: code = NotFound desc = could not find container \"dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096\": container with ID starting with dc122340a8d4ee1acb64d83b4c8e966e23415a4cbf2631de51da80666d2de096 not found: ID does not exist" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.260812 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e000bd6c-9e10-4b1c-8989-9d08c8b7a705" (UID: "e000bd6c-9e10-4b1c-8989-9d08c8b7a705"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.308538 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.308579 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6m7c\" (UniqueName: \"kubernetes.io/projected/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-kube-api-access-p6m7c\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.308592 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e000bd6c-9e10-4b1c-8989-9d08c8b7a705-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.455849 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zxg6h"] Nov 24 14:22:46 crc kubenswrapper[4822]: I1124 14:22:46.460418 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zxg6h"] Nov 24 14:22:47 crc kubenswrapper[4822]: I1124 14:22:47.303238 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:47 crc kubenswrapper[4822]: I1124 14:22:47.348121 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:22:47 crc kubenswrapper[4822]: I1124 14:22:47.362768 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:47 crc kubenswrapper[4822]: I1124 14:22:47.413801 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:47 crc kubenswrapper[4822]: I1124 14:22:47.710357 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" path="/var/lib/kubelet/pods/a80db4ed-d4de-49f6-95d7-513d109ba3e4/volumes" Nov 24 14:22:47 crc kubenswrapper[4822]: I1124 14:22:47.710934 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" path="/var/lib/kubelet/pods/e000bd6c-9e10-4b1c-8989-9d08c8b7a705/volumes" Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.168123 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wc4g"] Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.168385 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5wc4g" podUID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerName="registry-server" containerID="cri-o://e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6" gracePeriod=2 Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.503631 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.637194 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-catalog-content\") pod \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.637324 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-utilities\") pod \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.637364 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl7xs\" (UniqueName: \"kubernetes.io/projected/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-kube-api-access-vl7xs\") pod \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\" (UID: \"803cab3b-a4e9-4f87-a0a6-d3aa915c14de\") " Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.638334 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-utilities" (OuterVolumeSpecName: "utilities") pod "803cab3b-a4e9-4f87-a0a6-d3aa915c14de" (UID: "803cab3b-a4e9-4f87-a0a6-d3aa915c14de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.646498 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-kube-api-access-vl7xs" (OuterVolumeSpecName: "kube-api-access-vl7xs") pod "803cab3b-a4e9-4f87-a0a6-d3aa915c14de" (UID: "803cab3b-a4e9-4f87-a0a6-d3aa915c14de"). InnerVolumeSpecName "kube-api-access-vl7xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.661385 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "803cab3b-a4e9-4f87-a0a6-d3aa915c14de" (UID: "803cab3b-a4e9-4f87-a0a6-d3aa915c14de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.738910 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.738944 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:48 crc kubenswrapper[4822]: I1124 14:22:48.738954 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl7xs\" (UniqueName: \"kubernetes.io/projected/803cab3b-a4e9-4f87-a0a6-d3aa915c14de-kube-api-access-vl7xs\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.161379 4822 generic.go:334] "Generic (PLEG): container finished" podID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerID="e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6" exitCode=0 Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.161473 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5wc4g" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.161475 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wc4g" event={"ID":"803cab3b-a4e9-4f87-a0a6-d3aa915c14de","Type":"ContainerDied","Data":"e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6"} Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.161567 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wc4g" event={"ID":"803cab3b-a4e9-4f87-a0a6-d3aa915c14de","Type":"ContainerDied","Data":"cb088ca0d551ffa1ddec6e17829096e43383fecc0798afa126cee720aa6c74be"} Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.161614 4822 scope.go:117] "RemoveContainer" containerID="e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.183626 4822 scope.go:117] "RemoveContainer" containerID="25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.213401 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wc4g"] Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.213577 4822 scope.go:117] "RemoveContainer" containerID="b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.225398 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wc4g"] Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.242711 4822 scope.go:117] "RemoveContainer" containerID="e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6" Nov 24 14:22:49 crc kubenswrapper[4822]: E1124 14:22:49.243425 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6\": container with ID starting with e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6 not found: ID does not exist" containerID="e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.243512 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6"} err="failed to get container status \"e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6\": rpc error: code = NotFound desc = could not find container \"e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6\": container with ID starting with e9ed2be1f770991f9560c62c2c0c4c6f1d5769a8dc2ce2f24ccb7e8b546a09f6 not found: ID does not exist" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.243541 4822 scope.go:117] "RemoveContainer" containerID="25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2" Nov 24 14:22:49 crc kubenswrapper[4822]: E1124 14:22:49.244026 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2\": container with ID starting with 25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2 not found: ID does not exist" containerID="25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.244107 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2"} err="failed to get container status \"25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2\": rpc error: code = NotFound desc = could not find container \"25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2\": container with ID starting with 25e9f50e6b1bef5513263c2cca94bb2a7d6404ba056ad591018a20c23be788c2 not found: ID does not exist" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.244127 4822 scope.go:117] "RemoveContainer" containerID="b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166" Nov 24 14:22:49 crc kubenswrapper[4822]: E1124 14:22:49.244684 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166\": container with ID starting with b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166 not found: ID does not exist" containerID="b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.244726 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166"} err="failed to get container status \"b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166\": rpc error: code = NotFound desc = could not find container \"b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166\": container with ID starting with b5e87e7bff91a7492bdff3f0d7cab63cd0ee0cd07893aa1cbf693b779cd7f166 not found: ID does not exist" Nov 24 14:22:49 crc kubenswrapper[4822]: I1124 14:22:49.711088 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" path="/var/lib/kubelet/pods/803cab3b-a4e9-4f87-a0a6-d3aa915c14de/volumes" Nov 24 14:22:51 crc kubenswrapper[4822]: I1124 14:22:51.567389 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kkgmq"] Nov 24 14:22:51 crc kubenswrapper[4822]: I1124 14:22:51.567871 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kkgmq" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerName="registry-server" containerID="cri-o://2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c" gracePeriod=2 Nov 24 14:22:51 crc kubenswrapper[4822]: I1124 14:22:51.954413 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.081414 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc624\" (UniqueName: \"kubernetes.io/projected/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-kube-api-access-rc624\") pod \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.081861 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-catalog-content\") pod \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.081954 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-utilities\") pod \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\" (UID: \"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2\") " Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.082793 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-utilities" (OuterVolumeSpecName: "utilities") pod "d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" (UID: "d14f6b2d-cb93-4fd0-acde-a1d42a0179a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.089381 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-kube-api-access-rc624" (OuterVolumeSpecName: "kube-api-access-rc624") pod "d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" (UID: "d14f6b2d-cb93-4fd0-acde-a1d42a0179a2"). InnerVolumeSpecName "kube-api-access-rc624". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.181821 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" (UID: "d14f6b2d-cb93-4fd0-acde-a1d42a0179a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.182895 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.182934 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.182946 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc624\" (UniqueName: \"kubernetes.io/projected/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2-kube-api-access-rc624\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.186475 4822 generic.go:334] "Generic (PLEG): container finished" podID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerID="2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c" exitCode=0 Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.186533 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkgmq" event={"ID":"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2","Type":"ContainerDied","Data":"2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c"} Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.186581 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkgmq" event={"ID":"d14f6b2d-cb93-4fd0-acde-a1d42a0179a2","Type":"ContainerDied","Data":"61a5fd793de702c19f2a880a3fe8038714808d228f66c31931e8c3370823b40d"} Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.186610 4822 scope.go:117] "RemoveContainer" containerID="2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.186765 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkgmq" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.212617 4822 scope.go:117] "RemoveContainer" containerID="9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.232486 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kkgmq"] Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.236653 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kkgmq"] Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.250898 4822 scope.go:117] "RemoveContainer" containerID="a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.269874 4822 scope.go:117] "RemoveContainer" containerID="2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c" Nov 24 14:22:52 crc kubenswrapper[4822]: E1124 14:22:52.270941 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c\": container with ID starting with 2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c not found: ID does not exist" containerID="2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.271008 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c"} err="failed to get container status \"2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c\": rpc error: code = NotFound desc = could not find container \"2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c\": container with ID starting with 2a0d6da63618ae673b9e2551962046fa515a8b48d3780a043e06b0179d09195c not found: ID does not exist" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.271050 4822 scope.go:117] "RemoveContainer" containerID="9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e" Nov 24 14:22:52 crc kubenswrapper[4822]: E1124 14:22:52.271766 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e\": container with ID starting with 9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e not found: ID does not exist" containerID="9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.271908 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e"} err="failed to get container status \"9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e\": rpc error: code = NotFound desc = could not find container \"9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e\": container with ID starting with 9f73ea8f8cbff9040b1375b20b21fe5d9dea0fd460e1f959fde478bcfaa8eb0e not found: ID does not exist" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.272049 4822 scope.go:117] "RemoveContainer" containerID="a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032" Nov 24 14:22:52 crc kubenswrapper[4822]: E1124 14:22:52.272746 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032\": container with ID starting with a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032 not found: ID does not exist" containerID="a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032" Nov 24 14:22:52 crc kubenswrapper[4822]: I1124 14:22:52.272779 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032"} err="failed to get container status \"a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032\": rpc error: code = NotFound desc = could not find container \"a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032\": container with ID starting with a058b86232aa7e46701969b9e73eafe4fe93607f44938fbaecf7192891161032 not found: ID does not exist" Nov 24 14:22:53 crc kubenswrapper[4822]: I1124 14:22:53.712066 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" path="/var/lib/kubelet/pods/d14f6b2d-cb93-4fd0-acde-a1d42a0179a2/volumes" Nov 24 14:22:59 crc kubenswrapper[4822]: I1124 14:22:59.772001 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" podUID="902bd164-754e-4c87-b042-ca87c15d0263" containerName="oauth-openshift" containerID="cri-o://37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c" gracePeriod=15 Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.111900 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182385 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-session\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182455 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-trusted-ca-bundle\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182488 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-login\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182513 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-error\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182536 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-ocp-branding-template\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182566 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-router-certs\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182593 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-audit-policies\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182614 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-provider-selection\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182639 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-service-ca\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182665 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-idp-0-file-data\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182716 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-serving-cert\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182748 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgd7p\" (UniqueName: \"kubernetes.io/projected/902bd164-754e-4c87-b042-ca87c15d0263-kube-api-access-cgd7p\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182772 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/902bd164-754e-4c87-b042-ca87c15d0263-audit-dir\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.182821 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-cliconfig\") pod \"902bd164-754e-4c87-b042-ca87c15d0263\" (UID: \"902bd164-754e-4c87-b042-ca87c15d0263\") " Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.183698 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.184072 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.184079 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902bd164-754e-4c87-b042-ca87c15d0263-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.184607 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.184973 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.190176 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.190570 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.201353 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/902bd164-754e-4c87-b042-ca87c15d0263-kube-api-access-cgd7p" (OuterVolumeSpecName: "kube-api-access-cgd7p") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "kube-api-access-cgd7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.201646 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.202782 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.203570 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.203964 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.204625 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.204825 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "902bd164-754e-4c87-b042-ca87c15d0263" (UID: "902bd164-754e-4c87-b042-ca87c15d0263"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.239241 4822 generic.go:334] "Generic (PLEG): container finished" podID="902bd164-754e-4c87-b042-ca87c15d0263" containerID="37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c" exitCode=0 Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.239337 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.239352 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" event={"ID":"902bd164-754e-4c87-b042-ca87c15d0263","Type":"ContainerDied","Data":"37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c"} Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.239405 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gzxq" event={"ID":"902bd164-754e-4c87-b042-ca87c15d0263","Type":"ContainerDied","Data":"0ad157597f9cf9931ec4ba2591d5112aad1651dc0e1fadba8465e4851a4d282f"} Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.239436 4822 scope.go:117] "RemoveContainer" containerID="37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.266929 4822 scope.go:117] "RemoveContainer" containerID="37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.268214 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c\": container with ID starting with 37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c not found: ID does not exist" containerID="37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.268263 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c"} err="failed to get container status \"37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c\": rpc error: code = NotFound desc = could not find container \"37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c\": container with ID starting with 37301c205eea6a5c3b8de8f06c3b22516c6b32c48435ae93749d939e6ef45a0c not found: ID does not exist" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.280497 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gzxq"] Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.282885 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gzxq"] Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.285593 4822 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.285846 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.285941 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286014 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286079 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286135 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgd7p\" (UniqueName: \"kubernetes.io/projected/902bd164-754e-4c87-b042-ca87c15d0263-kube-api-access-cgd7p\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286215 4822 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/902bd164-754e-4c87-b042-ca87c15d0263-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286276 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286346 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286409 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286515 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286666 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286731 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.286788 4822 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/902bd164-754e-4c87-b042-ca87c15d0263-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.310903 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd"] Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311192 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311224 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311244 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902bd164-754e-4c87-b042-ca87c15d0263" containerName="oauth-openshift" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311254 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="902bd164-754e-4c87-b042-ca87c15d0263" containerName="oauth-openshift" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311265 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerName="extract-utilities" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311273 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerName="extract-utilities" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311282 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerName="extract-content" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311289 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerName="extract-content" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311300 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerName="extract-content" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311307 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerName="extract-content" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311321 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311328 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311341 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerName="extract-content" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311349 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerName="extract-content" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311359 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311366 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311381 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerName="extract-utilities" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311391 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerName="extract-utilities" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311403 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerName="extract-content" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311412 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerName="extract-content" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311425 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerName="extract-utilities" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311434 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerName="extract-utilities" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311443 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcca393b-facc-4010-95b7-9b1a0315a448" containerName="pruner" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311451 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcca393b-facc-4010-95b7-9b1a0315a448" containerName="pruner" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311463 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311471 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: E1124 14:23:00.311481 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerName="extract-utilities" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311489 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerName="extract-utilities" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311610 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="902bd164-754e-4c87-b042-ca87c15d0263" containerName="oauth-openshift" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311624 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a80db4ed-d4de-49f6-95d7-513d109ba3e4" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311637 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="803cab3b-a4e9-4f87-a0a6-d3aa915c14de" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311651 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="d14f6b2d-cb93-4fd0-acde-a1d42a0179a2" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311660 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcca393b-facc-4010-95b7-9b1a0315a448" containerName="pruner" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.311673 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="e000bd6c-9e10-4b1c-8989-9d08c8b7a705" containerName="registry-server" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.312171 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.314671 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.315160 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.315978 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.317717 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.319472 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.319801 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.320409 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.320806 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.321039 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.321810 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.322128 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.322240 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.327268 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.333610 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd"] Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.342515 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.357373 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.387948 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-template-login\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388012 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388037 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-audit-policies\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388062 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388084 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-service-ca\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388105 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388121 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50c96d31-86a6-4c52-b683-3143cf8f52a8-audit-dir\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388138 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-router-certs\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388161 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388177 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-session\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388226 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-template-error\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388244 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tccg\" (UniqueName: \"kubernetes.io/projected/50c96d31-86a6-4c52-b683-3143cf8f52a8-kube-api-access-9tccg\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388268 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.388286 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489276 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-service-ca\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489348 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489375 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50c96d31-86a6-4c52-b683-3143cf8f52a8-audit-dir\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489402 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-router-certs\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489434 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489463 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-session\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489508 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-template-error\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489534 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tccg\" (UniqueName: \"kubernetes.io/projected/50c96d31-86a6-4c52-b683-3143cf8f52a8-kube-api-access-9tccg\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489528 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50c96d31-86a6-4c52-b683-3143cf8f52a8-audit-dir\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489561 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489627 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489715 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-template-login\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489792 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489823 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-audit-policies\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.489863 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.490922 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.490926 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-audit-policies\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.490945 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.491069 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-service-ca\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.494087 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-template-error\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.494295 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-router-certs\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.494379 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.494635 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.496316 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.497609 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-template-login\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.498940 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-system-session\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.501453 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/50c96d31-86a6-4c52-b683-3143cf8f52a8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.512398 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tccg\" (UniqueName: \"kubernetes.io/projected/50c96d31-86a6-4c52-b683-3143cf8f52a8-kube-api-access-9tccg\") pod \"oauth-openshift-6775b6d8cc-c2tvd\" (UID: \"50c96d31-86a6-4c52-b683-3143cf8f52a8\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.649756 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:00 crc kubenswrapper[4822]: I1124 14:23:00.897738 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd"] Nov 24 14:23:01 crc kubenswrapper[4822]: I1124 14:23:01.250772 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" event={"ID":"50c96d31-86a6-4c52-b683-3143cf8f52a8","Type":"ContainerStarted","Data":"edc8bf34eafcfc23b7e15e49946efee62576ec45c45605025fa63e8edc677c25"} Nov 24 14:23:01 crc kubenswrapper[4822]: I1124 14:23:01.251078 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:01 crc kubenswrapper[4822]: I1124 14:23:01.251090 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" event={"ID":"50c96d31-86a6-4c52-b683-3143cf8f52a8","Type":"ContainerStarted","Data":"3768b9a610d01b224cc266d04377e01f11789a5da6d4dbd2ea8d5f54efe552e0"} Nov 24 14:23:01 crc kubenswrapper[4822]: I1124 14:23:01.279075 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" podStartSLOduration=27.27903347 podStartE2EDuration="27.27903347s" podCreationTimestamp="2025-11-24 14:22:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:23:01.277707678 +0000 UTC m=+218.394348175" watchObservedRunningTime="2025-11-24 14:23:01.27903347 +0000 UTC m=+218.395673967" Nov 24 14:23:01 crc kubenswrapper[4822]: I1124 14:23:01.715004 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="902bd164-754e-4c87-b042-ca87c15d0263" path="/var/lib/kubelet/pods/902bd164-754e-4c87-b042-ca87c15d0263/volumes" Nov 24 14:23:01 crc kubenswrapper[4822]: I1124 14:23:01.739310 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6775b6d8cc-c2tvd" Nov 24 14:23:11 crc kubenswrapper[4822]: I1124 14:23:11.578359 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:23:11 crc kubenswrapper[4822]: I1124 14:23:11.579857 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:23:11 crc kubenswrapper[4822]: I1124 14:23:11.579966 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:23:11 crc kubenswrapper[4822]: I1124 14:23:11.581167 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:23:11 crc kubenswrapper[4822]: I1124 14:23:11.581317 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f" gracePeriod=600 Nov 24 14:23:12 crc kubenswrapper[4822]: I1124 14:23:12.318584 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f" exitCode=0 Nov 24 14:23:12 crc kubenswrapper[4822]: I1124 14:23:12.319158 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f"} Nov 24 14:23:12 crc kubenswrapper[4822]: I1124 14:23:12.319190 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"7115168ad50bfd88ed09dd9295ac9935a4ada79bd3272f63430ce467dc596e9b"} Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.724384 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xjpf5"] Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.725315 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xjpf5" podUID="b002a922-0295-4465-a424-21c07d16a661" containerName="registry-server" containerID="cri-o://67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d" gracePeriod=30 Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.739437 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhbbs"] Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.739711 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jhbbs" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerName="registry-server" containerID="cri-o://9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2" gracePeriod=30 Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.747794 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-95knp"] Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.748068 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" podUID="f54974f4-d4d0-4962-baf5-d145bd8fc5f9" containerName="marketplace-operator" containerID="cri-o://8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e" gracePeriod=30 Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.759791 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lgxb4"] Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.760118 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lgxb4" podUID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerName="registry-server" containerID="cri-o://2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934" gracePeriod=30 Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.771965 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xndhl"] Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.774113 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.781583 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x55mp"] Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.782076 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x55mp" podUID="b17ebf75-0076-48f5-898f-0196a270a555" containerName="registry-server" containerID="cri-o://99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf" gracePeriod=30 Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.809597 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xndhl"] Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.877826 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1d96a94a-717c-42c8-90e2-ae98a6edcb99-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xndhl\" (UID: \"1d96a94a-717c-42c8-90e2-ae98a6edcb99\") " pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.877886 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdq8b\" (UniqueName: \"kubernetes.io/projected/1d96a94a-717c-42c8-90e2-ae98a6edcb99-kube-api-access-fdq8b\") pod \"marketplace-operator-79b997595-xndhl\" (UID: \"1d96a94a-717c-42c8-90e2-ae98a6edcb99\") " pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.877931 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d96a94a-717c-42c8-90e2-ae98a6edcb99-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xndhl\" (UID: \"1d96a94a-717c-42c8-90e2-ae98a6edcb99\") " pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.979162 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdq8b\" (UniqueName: \"kubernetes.io/projected/1d96a94a-717c-42c8-90e2-ae98a6edcb99-kube-api-access-fdq8b\") pod \"marketplace-operator-79b997595-xndhl\" (UID: \"1d96a94a-717c-42c8-90e2-ae98a6edcb99\") " pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.979257 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d96a94a-717c-42c8-90e2-ae98a6edcb99-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xndhl\" (UID: \"1d96a94a-717c-42c8-90e2-ae98a6edcb99\") " pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.979320 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1d96a94a-717c-42c8-90e2-ae98a6edcb99-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xndhl\" (UID: \"1d96a94a-717c-42c8-90e2-ae98a6edcb99\") " pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.981908 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d96a94a-717c-42c8-90e2-ae98a6edcb99-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xndhl\" (UID: \"1d96a94a-717c-42c8-90e2-ae98a6edcb99\") " pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:13 crc kubenswrapper[4822]: I1124 14:23:13.986620 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1d96a94a-717c-42c8-90e2-ae98a6edcb99-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xndhl\" (UID: \"1d96a94a-717c-42c8-90e2-ae98a6edcb99\") " pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.000891 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdq8b\" (UniqueName: \"kubernetes.io/projected/1d96a94a-717c-42c8-90e2-ae98a6edcb99-kube-api-access-fdq8b\") pod \"marketplace-operator-79b997595-xndhl\" (UID: \"1d96a94a-717c-42c8-90e2-ae98a6edcb99\") " pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.143796 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.147474 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.157836 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.160754 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.163827 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.181378 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.282905 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-catalog-content\") pod \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.282947 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p6rv\" (UniqueName: \"kubernetes.io/projected/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-kube-api-access-8p6rv\") pod \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.282980 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-utilities\") pod \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283010 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-operator-metrics\") pod \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283034 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6zwv\" (UniqueName: \"kubernetes.io/projected/decf7f1e-5ed6-4026-ad3a-41144a216fd0-kube-api-access-n6zwv\") pod \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283051 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b26fl\" (UniqueName: \"kubernetes.io/projected/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-kube-api-access-b26fl\") pod \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283092 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-catalog-content\") pod \"b002a922-0295-4465-a424-21c07d16a661\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283107 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-utilities\") pod \"b17ebf75-0076-48f5-898f-0196a270a555\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283131 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-utilities\") pod \"b002a922-0295-4465-a424-21c07d16a661\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283170 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-trusted-ca\") pod \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\" (UID: \"f54974f4-d4d0-4962-baf5-d145bd8fc5f9\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283184 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-catalog-content\") pod \"b17ebf75-0076-48f5-898f-0196a270a555\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283217 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-catalog-content\") pod \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\" (UID: \"decf7f1e-5ed6-4026-ad3a-41144a216fd0\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283236 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7p5k\" (UniqueName: \"kubernetes.io/projected/b002a922-0295-4465-a424-21c07d16a661-kube-api-access-j7p5k\") pod \"b002a922-0295-4465-a424-21c07d16a661\" (UID: \"b002a922-0295-4465-a424-21c07d16a661\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283257 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxbtj\" (UniqueName: \"kubernetes.io/projected/b17ebf75-0076-48f5-898f-0196a270a555-kube-api-access-lxbtj\") pod \"b17ebf75-0076-48f5-898f-0196a270a555\" (UID: \"b17ebf75-0076-48f5-898f-0196a270a555\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.283275 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-utilities\") pod \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\" (UID: \"5e5e4406-043a-46a3-8bf8-9de3cdbae55c\") " Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.285599 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-utilities" (OuterVolumeSpecName: "utilities") pod "b002a922-0295-4465-a424-21c07d16a661" (UID: "b002a922-0295-4465-a424-21c07d16a661"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.289665 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b002a922-0295-4465-a424-21c07d16a661-kube-api-access-j7p5k" (OuterVolumeSpecName: "kube-api-access-j7p5k") pod "b002a922-0295-4465-a424-21c07d16a661" (UID: "b002a922-0295-4465-a424-21c07d16a661"). InnerVolumeSpecName "kube-api-access-j7p5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.290446 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f54974f4-d4d0-4962-baf5-d145bd8fc5f9" (UID: "f54974f4-d4d0-4962-baf5-d145bd8fc5f9"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.290541 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-kube-api-access-8p6rv" (OuterVolumeSpecName: "kube-api-access-8p6rv") pod "f54974f4-d4d0-4962-baf5-d145bd8fc5f9" (UID: "f54974f4-d4d0-4962-baf5-d145bd8fc5f9"). InnerVolumeSpecName "kube-api-access-8p6rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.294399 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-utilities" (OuterVolumeSpecName: "utilities") pod "decf7f1e-5ed6-4026-ad3a-41144a216fd0" (UID: "decf7f1e-5ed6-4026-ad3a-41144a216fd0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.294649 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-utilities" (OuterVolumeSpecName: "utilities") pod "b17ebf75-0076-48f5-898f-0196a270a555" (UID: "b17ebf75-0076-48f5-898f-0196a270a555"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.294722 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-utilities" (OuterVolumeSpecName: "utilities") pod "5e5e4406-043a-46a3-8bf8-9de3cdbae55c" (UID: "5e5e4406-043a-46a3-8bf8-9de3cdbae55c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.299601 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/decf7f1e-5ed6-4026-ad3a-41144a216fd0-kube-api-access-n6zwv" (OuterVolumeSpecName: "kube-api-access-n6zwv") pod "decf7f1e-5ed6-4026-ad3a-41144a216fd0" (UID: "decf7f1e-5ed6-4026-ad3a-41144a216fd0"). InnerVolumeSpecName "kube-api-access-n6zwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.301350 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b17ebf75-0076-48f5-898f-0196a270a555-kube-api-access-lxbtj" (OuterVolumeSpecName: "kube-api-access-lxbtj") pod "b17ebf75-0076-48f5-898f-0196a270a555" (UID: "b17ebf75-0076-48f5-898f-0196a270a555"). InnerVolumeSpecName "kube-api-access-lxbtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.304488 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f54974f4-d4d0-4962-baf5-d145bd8fc5f9" (UID: "f54974f4-d4d0-4962-baf5-d145bd8fc5f9"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.308813 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-kube-api-access-b26fl" (OuterVolumeSpecName: "kube-api-access-b26fl") pod "5e5e4406-043a-46a3-8bf8-9de3cdbae55c" (UID: "5e5e4406-043a-46a3-8bf8-9de3cdbae55c"). InnerVolumeSpecName "kube-api-access-b26fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.314379 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e5e4406-043a-46a3-8bf8-9de3cdbae55c" (UID: "5e5e4406-043a-46a3-8bf8-9de3cdbae55c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.338197 4822 generic.go:334] "Generic (PLEG): container finished" podID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerID="2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934" exitCode=0 Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.338324 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lgxb4" event={"ID":"5e5e4406-043a-46a3-8bf8-9de3cdbae55c","Type":"ContainerDied","Data":"2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.338356 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lgxb4" event={"ID":"5e5e4406-043a-46a3-8bf8-9de3cdbae55c","Type":"ContainerDied","Data":"a8abda4e7f945c92a1ea3611a917efc2cb5064d4199660a12f60aaf846d35383"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.338376 4822 scope.go:117] "RemoveContainer" containerID="2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.338539 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lgxb4" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.342649 4822 generic.go:334] "Generic (PLEG): container finished" podID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerID="9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2" exitCode=0 Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.342708 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbbs" event={"ID":"decf7f1e-5ed6-4026-ad3a-41144a216fd0","Type":"ContainerDied","Data":"9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.342735 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbbs" event={"ID":"decf7f1e-5ed6-4026-ad3a-41144a216fd0","Type":"ContainerDied","Data":"94b718c1a2d2e7fa3605d97e60de80b8e961c7237f5f274648a7623e819237e4"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.342820 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhbbs" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.346215 4822 generic.go:334] "Generic (PLEG): container finished" podID="b17ebf75-0076-48f5-898f-0196a270a555" containerID="99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf" exitCode=0 Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.346224 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55mp" event={"ID":"b17ebf75-0076-48f5-898f-0196a270a555","Type":"ContainerDied","Data":"99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.346284 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x55mp" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.346303 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55mp" event={"ID":"b17ebf75-0076-48f5-898f-0196a270a555","Type":"ContainerDied","Data":"acd3be21cebe24df5ef922fa97f3f6bf5b765628babf94ddc1c355459393d256"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.347381 4822 generic.go:334] "Generic (PLEG): container finished" podID="f54974f4-d4d0-4962-baf5-d145bd8fc5f9" containerID="8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e" exitCode=0 Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.347474 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" event={"ID":"f54974f4-d4d0-4962-baf5-d145bd8fc5f9","Type":"ContainerDied","Data":"8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.347492 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" event={"ID":"f54974f4-d4d0-4962-baf5-d145bd8fc5f9","Type":"ContainerDied","Data":"316a77d30678ea484470c8402bf7857ba3b3d927a0600dbefbaf85f88f85d532"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.347539 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-95knp" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.351261 4822 generic.go:334] "Generic (PLEG): container finished" podID="b002a922-0295-4465-a424-21c07d16a661" containerID="67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d" exitCode=0 Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.351306 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjpf5" event={"ID":"b002a922-0295-4465-a424-21c07d16a661","Type":"ContainerDied","Data":"67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.351328 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjpf5" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.351336 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjpf5" event={"ID":"b002a922-0295-4465-a424-21c07d16a661","Type":"ContainerDied","Data":"2fc1dd559460aa3270b4f95a9a9c75daf740fe684768aff3ffbe5cd183aeb71d"} Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.356195 4822 scope.go:117] "RemoveContainer" containerID="d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.365743 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "decf7f1e-5ed6-4026-ad3a-41144a216fd0" (UID: "decf7f1e-5ed6-4026-ad3a-41144a216fd0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.374057 4822 scope.go:117] "RemoveContainer" containerID="263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.377159 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b002a922-0295-4465-a424-21c07d16a661" (UID: "b002a922-0295-4465-a424-21c07d16a661"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385854 4822 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385880 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385889 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7p5k\" (UniqueName: \"kubernetes.io/projected/b002a922-0295-4465-a424-21c07d16a661-kube-api-access-j7p5k\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385898 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxbtj\" (UniqueName: \"kubernetes.io/projected/b17ebf75-0076-48f5-898f-0196a270a555-kube-api-access-lxbtj\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385908 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385917 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385925 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p6rv\" (UniqueName: \"kubernetes.io/projected/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-kube-api-access-8p6rv\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385934 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/decf7f1e-5ed6-4026-ad3a-41144a216fd0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385942 4822 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f54974f4-d4d0-4962-baf5-d145bd8fc5f9-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385950 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6zwv\" (UniqueName: \"kubernetes.io/projected/decf7f1e-5ed6-4026-ad3a-41144a216fd0-kube-api-access-n6zwv\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385959 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b26fl\" (UniqueName: \"kubernetes.io/projected/5e5e4406-043a-46a3-8bf8-9de3cdbae55c-kube-api-access-b26fl\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385969 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385977 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.385985 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b002a922-0295-4465-a424-21c07d16a661-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.392399 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-95knp"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.399474 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-95knp"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.400601 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xndhl"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.403052 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lgxb4"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.406038 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lgxb4"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.414243 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b17ebf75-0076-48f5-898f-0196a270a555" (UID: "b17ebf75-0076-48f5-898f-0196a270a555"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.420755 4822 scope.go:117] "RemoveContainer" containerID="2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.422795 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934\": container with ID starting with 2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934 not found: ID does not exist" containerID="2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.423071 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934"} err="failed to get container status \"2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934\": rpc error: code = NotFound desc = could not find container \"2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934\": container with ID starting with 2784a1da7aa303592749e1e7ce7cf7b715d686a573949569547dc7b5818fb934 not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.423106 4822 scope.go:117] "RemoveContainer" containerID="d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.423380 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40\": container with ID starting with d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40 not found: ID does not exist" containerID="d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.423401 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40"} err="failed to get container status \"d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40\": rpc error: code = NotFound desc = could not find container \"d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40\": container with ID starting with d97301888327d2592b41f36dd087d8876b1745c1acbd4727ed91fa75696b9a40 not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.423416 4822 scope.go:117] "RemoveContainer" containerID="263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.423574 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd\": container with ID starting with 263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd not found: ID does not exist" containerID="263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.423596 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd"} err="failed to get container status \"263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd\": rpc error: code = NotFound desc = could not find container \"263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd\": container with ID starting with 263073068e95c97b2420848c68787491497b53c362b53f81a1c73be2cb7866bd not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.423608 4822 scope.go:117] "RemoveContainer" containerID="9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.438999 4822 scope.go:117] "RemoveContainer" containerID="10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.456672 4822 scope.go:117] "RemoveContainer" containerID="b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.469806 4822 scope.go:117] "RemoveContainer" containerID="9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.470406 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2\": container with ID starting with 9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2 not found: ID does not exist" containerID="9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.470449 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2"} err="failed to get container status \"9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2\": rpc error: code = NotFound desc = could not find container \"9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2\": container with ID starting with 9bfb0936f0221741c6070d82db65fd5e95107b2790faf87de24f230727db7be2 not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.470474 4822 scope.go:117] "RemoveContainer" containerID="10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.470823 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab\": container with ID starting with 10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab not found: ID does not exist" containerID="10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.470901 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab"} err="failed to get container status \"10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab\": rpc error: code = NotFound desc = could not find container \"10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab\": container with ID starting with 10dbd99ed483309cb528257e843f1d6b55bc4478d2575b0881c56943250f39ab not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.470972 4822 scope.go:117] "RemoveContainer" containerID="b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.471281 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6\": container with ID starting with b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6 not found: ID does not exist" containerID="b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.471372 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6"} err="failed to get container status \"b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6\": rpc error: code = NotFound desc = could not find container \"b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6\": container with ID starting with b0e4c3ae80f64bfc337e8e10904b618882807b9a694d3e14c545607358a4a3d6 not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.471436 4822 scope.go:117] "RemoveContainer" containerID="99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.487065 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17ebf75-0076-48f5-898f-0196a270a555-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.492426 4822 scope.go:117] "RemoveContainer" containerID="1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.524853 4822 scope.go:117] "RemoveContainer" containerID="cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.538173 4822 scope.go:117] "RemoveContainer" containerID="99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.538898 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf\": container with ID starting with 99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf not found: ID does not exist" containerID="99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.538957 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf"} err="failed to get container status \"99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf\": rpc error: code = NotFound desc = could not find container \"99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf\": container with ID starting with 99dc12aa250419e0d4b25a5a7f5227a1958f5e21c1f0cd56c91d593b10ad73bf not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.538996 4822 scope.go:117] "RemoveContainer" containerID="1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.539510 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8\": container with ID starting with 1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8 not found: ID does not exist" containerID="1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.539824 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8"} err="failed to get container status \"1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8\": rpc error: code = NotFound desc = could not find container \"1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8\": container with ID starting with 1f1d890aa7b7ba136e5b9a966982277c0ab11753bb6255e0414ae39024394bc8 not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.539857 4822 scope.go:117] "RemoveContainer" containerID="cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.540233 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e\": container with ID starting with cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e not found: ID does not exist" containerID="cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.540254 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e"} err="failed to get container status \"cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e\": rpc error: code = NotFound desc = could not find container \"cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e\": container with ID starting with cc0c4d8f6711b49d38448eeb0ec4b64998fa54d31f72bd958c4eb18da666709e not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.540285 4822 scope.go:117] "RemoveContainer" containerID="8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.552310 4822 scope.go:117] "RemoveContainer" containerID="8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.552750 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e\": container with ID starting with 8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e not found: ID does not exist" containerID="8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.552795 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e"} err="failed to get container status \"8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e\": rpc error: code = NotFound desc = could not find container \"8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e\": container with ID starting with 8394eb01312d1583d53a2af4e711705064bb99c241ba2425a6b43c001dddf62e not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.552830 4822 scope.go:117] "RemoveContainer" containerID="67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.565420 4822 scope.go:117] "RemoveContainer" containerID="25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.578995 4822 scope.go:117] "RemoveContainer" containerID="40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.596960 4822 scope.go:117] "RemoveContainer" containerID="67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.597465 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d\": container with ID starting with 67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d not found: ID does not exist" containerID="67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.597501 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d"} err="failed to get container status \"67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d\": rpc error: code = NotFound desc = could not find container \"67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d\": container with ID starting with 67d48df828a7fed5a40c48d32bd59c8adbdbc936a4338857c6806649a8b2816d not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.597525 4822 scope.go:117] "RemoveContainer" containerID="25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.597929 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5\": container with ID starting with 25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5 not found: ID does not exist" containerID="25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.597950 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5"} err="failed to get container status \"25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5\": rpc error: code = NotFound desc = could not find container \"25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5\": container with ID starting with 25aab571e542229d4bd9f92935c1e4b4bf4eba40c9c0d0d1c4e096ff9a6e33b5 not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.597967 4822 scope.go:117] "RemoveContainer" containerID="40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7" Nov 24 14:23:14 crc kubenswrapper[4822]: E1124 14:23:14.598192 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7\": container with ID starting with 40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7 not found: ID does not exist" containerID="40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.598309 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7"} err="failed to get container status \"40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7\": rpc error: code = NotFound desc = could not find container \"40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7\": container with ID starting with 40fd419a3851454fd4dc816ceac2d5b5a6da3736186bc6f5911120271f35a8b7 not found: ID does not exist" Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.683606 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhbbs"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.688130 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jhbbs"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.699225 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xjpf5"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.712182 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xjpf5"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.715015 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x55mp"] Nov 24 14:23:14 crc kubenswrapper[4822]: I1124 14:23:14.717009 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x55mp"] Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.363137 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" event={"ID":"1d96a94a-717c-42c8-90e2-ae98a6edcb99","Type":"ContainerStarted","Data":"28118f8c67fa946e5017cef8d8a9d4b5abe16dd11b04325fe3d0a10465daed63"} Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.367269 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" event={"ID":"1d96a94a-717c-42c8-90e2-ae98a6edcb99","Type":"ContainerStarted","Data":"9fed122ef49f766a4a04d4989b6ffcbef593bf21d94a92338b2f351ad0d178b9"} Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.367589 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.368903 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.386719 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xndhl" podStartSLOduration=2.38669564 podStartE2EDuration="2.38669564s" podCreationTimestamp="2025-11-24 14:23:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:23:15.384346895 +0000 UTC m=+232.500987392" watchObservedRunningTime="2025-11-24 14:23:15.38669564 +0000 UTC m=+232.503336117" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.715499 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" path="/var/lib/kubelet/pods/5e5e4406-043a-46a3-8bf8-9de3cdbae55c/volumes" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.717252 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b002a922-0295-4465-a424-21c07d16a661" path="/var/lib/kubelet/pods/b002a922-0295-4465-a424-21c07d16a661/volumes" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.717981 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b17ebf75-0076-48f5-898f-0196a270a555" path="/var/lib/kubelet/pods/b17ebf75-0076-48f5-898f-0196a270a555/volumes" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.719700 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" path="/var/lib/kubelet/pods/decf7f1e-5ed6-4026-ad3a-41144a216fd0/volumes" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.721076 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f54974f4-d4d0-4962-baf5-d145bd8fc5f9" path="/var/lib/kubelet/pods/f54974f4-d4d0-4962-baf5-d145bd8fc5f9/volumes" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.939988 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-28rbg"] Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.940527 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b002a922-0295-4465-a424-21c07d16a661" containerName="extract-utilities" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.940614 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b002a922-0295-4465-a424-21c07d16a661" containerName="extract-utilities" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.940674 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b002a922-0295-4465-a424-21c07d16a661" containerName="extract-content" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.940753 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b002a922-0295-4465-a424-21c07d16a661" containerName="extract-content" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.940813 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.941225 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.941305 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17ebf75-0076-48f5-898f-0196a270a555" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.941358 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17ebf75-0076-48f5-898f-0196a270a555" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.941491 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerName="extract-utilities" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.941556 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerName="extract-utilities" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.941613 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerName="extract-utilities" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.941663 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerName="extract-utilities" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.941719 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17ebf75-0076-48f5-898f-0196a270a555" containerName="extract-content" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.941768 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17ebf75-0076-48f5-898f-0196a270a555" containerName="extract-content" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.941822 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.941873 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.941931 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b002a922-0295-4465-a424-21c07d16a661" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.941986 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b002a922-0295-4465-a424-21c07d16a661" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.942047 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f54974f4-d4d0-4962-baf5-d145bd8fc5f9" containerName="marketplace-operator" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.942102 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f54974f4-d4d0-4962-baf5-d145bd8fc5f9" containerName="marketplace-operator" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.942154 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerName="extract-content" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.942633 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerName="extract-content" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.942773 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17ebf75-0076-48f5-898f-0196a270a555" containerName="extract-utilities" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.942828 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17ebf75-0076-48f5-898f-0196a270a555" containerName="extract-utilities" Nov 24 14:23:15 crc kubenswrapper[4822]: E1124 14:23:15.942885 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerName="extract-content" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.942945 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerName="extract-content" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.943076 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="decf7f1e-5ed6-4026-ad3a-41144a216fd0" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.943135 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="b17ebf75-0076-48f5-898f-0196a270a555" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.943191 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5e4406-043a-46a3-8bf8-9de3cdbae55c" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.943264 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="f54974f4-d4d0-4962-baf5-d145bd8fc5f9" containerName="marketplace-operator" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.943337 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="b002a922-0295-4465-a424-21c07d16a661" containerName="registry-server" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.944098 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.949841 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 14:23:15 crc kubenswrapper[4822]: I1124 14:23:15.953406 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-28rbg"] Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.105533 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-catalog-content\") pod \"certified-operators-28rbg\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.105627 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-utilities\") pod \"certified-operators-28rbg\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.105665 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wttpp\" (UniqueName: \"kubernetes.io/projected/5c76e331-f8e8-45e9-9afb-2add7327ebd1-kube-api-access-wttpp\") pod \"certified-operators-28rbg\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.150274 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vlfkr"] Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.152041 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.156879 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.165613 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vlfkr"] Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.206421 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-catalog-content\") pod \"certified-operators-28rbg\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.206527 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-utilities\") pod \"certified-operators-28rbg\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.206573 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wttpp\" (UniqueName: \"kubernetes.io/projected/5c76e331-f8e8-45e9-9afb-2add7327ebd1-kube-api-access-wttpp\") pod \"certified-operators-28rbg\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.207470 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-utilities\") pod \"certified-operators-28rbg\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.207895 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-catalog-content\") pod \"certified-operators-28rbg\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.236373 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wttpp\" (UniqueName: \"kubernetes.io/projected/5c76e331-f8e8-45e9-9afb-2add7327ebd1-kube-api-access-wttpp\") pod \"certified-operators-28rbg\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.268359 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.307797 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv5kh\" (UniqueName: \"kubernetes.io/projected/e5fca0b7-7cab-4a35-a0ef-88a75d92707a-kube-api-access-fv5kh\") pod \"redhat-marketplace-vlfkr\" (UID: \"e5fca0b7-7cab-4a35-a0ef-88a75d92707a\") " pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.307906 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5fca0b7-7cab-4a35-a0ef-88a75d92707a-utilities\") pod \"redhat-marketplace-vlfkr\" (UID: \"e5fca0b7-7cab-4a35-a0ef-88a75d92707a\") " pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.307946 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5fca0b7-7cab-4a35-a0ef-88a75d92707a-catalog-content\") pod \"redhat-marketplace-vlfkr\" (UID: \"e5fca0b7-7cab-4a35-a0ef-88a75d92707a\") " pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.408970 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv5kh\" (UniqueName: \"kubernetes.io/projected/e5fca0b7-7cab-4a35-a0ef-88a75d92707a-kube-api-access-fv5kh\") pod \"redhat-marketplace-vlfkr\" (UID: \"e5fca0b7-7cab-4a35-a0ef-88a75d92707a\") " pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.411870 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5fca0b7-7cab-4a35-a0ef-88a75d92707a-utilities\") pod \"redhat-marketplace-vlfkr\" (UID: \"e5fca0b7-7cab-4a35-a0ef-88a75d92707a\") " pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.412712 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5fca0b7-7cab-4a35-a0ef-88a75d92707a-catalog-content\") pod \"redhat-marketplace-vlfkr\" (UID: \"e5fca0b7-7cab-4a35-a0ef-88a75d92707a\") " pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.413487 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5fca0b7-7cab-4a35-a0ef-88a75d92707a-utilities\") pod \"redhat-marketplace-vlfkr\" (UID: \"e5fca0b7-7cab-4a35-a0ef-88a75d92707a\") " pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.413499 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5fca0b7-7cab-4a35-a0ef-88a75d92707a-catalog-content\") pod \"redhat-marketplace-vlfkr\" (UID: \"e5fca0b7-7cab-4a35-a0ef-88a75d92707a\") " pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.431877 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv5kh\" (UniqueName: \"kubernetes.io/projected/e5fca0b7-7cab-4a35-a0ef-88a75d92707a-kube-api-access-fv5kh\") pod \"redhat-marketplace-vlfkr\" (UID: \"e5fca0b7-7cab-4a35-a0ef-88a75d92707a\") " pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.468538 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.498150 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-28rbg"] Nov 24 14:23:16 crc kubenswrapper[4822]: I1124 14:23:16.867366 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vlfkr"] Nov 24 14:23:16 crc kubenswrapper[4822]: W1124 14:23:16.873619 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5fca0b7_7cab_4a35_a0ef_88a75d92707a.slice/crio-df33fb168f4e679ca42ea8500ab65b10dd0680fbec9f023762f748b8fa3f7207 WatchSource:0}: Error finding container df33fb168f4e679ca42ea8500ab65b10dd0680fbec9f023762f748b8fa3f7207: Status 404 returned error can't find the container with id df33fb168f4e679ca42ea8500ab65b10dd0680fbec9f023762f748b8fa3f7207 Nov 24 14:23:17 crc kubenswrapper[4822]: I1124 14:23:17.375505 4822 generic.go:334] "Generic (PLEG): container finished" podID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerID="7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad" exitCode=0 Nov 24 14:23:17 crc kubenswrapper[4822]: I1124 14:23:17.375562 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28rbg" event={"ID":"5c76e331-f8e8-45e9-9afb-2add7327ebd1","Type":"ContainerDied","Data":"7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad"} Nov 24 14:23:17 crc kubenswrapper[4822]: I1124 14:23:17.375633 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28rbg" event={"ID":"5c76e331-f8e8-45e9-9afb-2add7327ebd1","Type":"ContainerStarted","Data":"905e8c024cd3af654c57166b4e5014d38a1d1530af4233882cbbe8d329ebb5a8"} Nov 24 14:23:17 crc kubenswrapper[4822]: I1124 14:23:17.381564 4822 generic.go:334] "Generic (PLEG): container finished" podID="e5fca0b7-7cab-4a35-a0ef-88a75d92707a" containerID="dbdf4965e0a8ed7c2c151a64745c2e981161457109b13b7b0e17a013bff9622a" exitCode=0 Nov 24 14:23:17 crc kubenswrapper[4822]: I1124 14:23:17.381649 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlfkr" event={"ID":"e5fca0b7-7cab-4a35-a0ef-88a75d92707a","Type":"ContainerDied","Data":"dbdf4965e0a8ed7c2c151a64745c2e981161457109b13b7b0e17a013bff9622a"} Nov 24 14:23:17 crc kubenswrapper[4822]: I1124 14:23:17.381687 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlfkr" event={"ID":"e5fca0b7-7cab-4a35-a0ef-88a75d92707a","Type":"ContainerStarted","Data":"df33fb168f4e679ca42ea8500ab65b10dd0680fbec9f023762f748b8fa3f7207"} Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.350129 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nvv4m"] Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.353233 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.356923 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.363441 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nvv4m"] Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.394599 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28rbg" event={"ID":"5c76e331-f8e8-45e9-9afb-2add7327ebd1","Type":"ContainerStarted","Data":"93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0"} Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.544913 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg26c\" (UniqueName: \"kubernetes.io/projected/39db4f3b-fa4e-41ad-92c9-5e682207c429-kube-api-access-hg26c\") pod \"redhat-operators-nvv4m\" (UID: \"39db4f3b-fa4e-41ad-92c9-5e682207c429\") " pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.545001 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39db4f3b-fa4e-41ad-92c9-5e682207c429-utilities\") pod \"redhat-operators-nvv4m\" (UID: \"39db4f3b-fa4e-41ad-92c9-5e682207c429\") " pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.545318 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39db4f3b-fa4e-41ad-92c9-5e682207c429-catalog-content\") pod \"redhat-operators-nvv4m\" (UID: \"39db4f3b-fa4e-41ad-92c9-5e682207c429\") " pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.546441 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9mzfn"] Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.547494 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.549373 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.563282 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9mzfn"] Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.647302 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-catalog-content\") pod \"community-operators-9mzfn\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.647365 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39db4f3b-fa4e-41ad-92c9-5e682207c429-catalog-content\") pod \"redhat-operators-nvv4m\" (UID: \"39db4f3b-fa4e-41ad-92c9-5e682207c429\") " pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.647416 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-utilities\") pod \"community-operators-9mzfn\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.647467 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg26c\" (UniqueName: \"kubernetes.io/projected/39db4f3b-fa4e-41ad-92c9-5e682207c429-kube-api-access-hg26c\") pod \"redhat-operators-nvv4m\" (UID: \"39db4f3b-fa4e-41ad-92c9-5e682207c429\") " pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.647532 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwkw5\" (UniqueName: \"kubernetes.io/projected/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-kube-api-access-mwkw5\") pod \"community-operators-9mzfn\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.647555 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39db4f3b-fa4e-41ad-92c9-5e682207c429-utilities\") pod \"redhat-operators-nvv4m\" (UID: \"39db4f3b-fa4e-41ad-92c9-5e682207c429\") " pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.647895 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39db4f3b-fa4e-41ad-92c9-5e682207c429-catalog-content\") pod \"redhat-operators-nvv4m\" (UID: \"39db4f3b-fa4e-41ad-92c9-5e682207c429\") " pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.648262 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39db4f3b-fa4e-41ad-92c9-5e682207c429-utilities\") pod \"redhat-operators-nvv4m\" (UID: \"39db4f3b-fa4e-41ad-92c9-5e682207c429\") " pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.668069 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg26c\" (UniqueName: \"kubernetes.io/projected/39db4f3b-fa4e-41ad-92c9-5e682207c429-kube-api-access-hg26c\") pod \"redhat-operators-nvv4m\" (UID: \"39db4f3b-fa4e-41ad-92c9-5e682207c429\") " pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.748355 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.749394 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwkw5\" (UniqueName: \"kubernetes.io/projected/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-kube-api-access-mwkw5\") pod \"community-operators-9mzfn\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.749430 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-catalog-content\") pod \"community-operators-9mzfn\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.749459 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-utilities\") pod \"community-operators-9mzfn\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.749847 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-utilities\") pod \"community-operators-9mzfn\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.749957 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-catalog-content\") pod \"community-operators-9mzfn\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.766667 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwkw5\" (UniqueName: \"kubernetes.io/projected/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-kube-api-access-mwkw5\") pod \"community-operators-9mzfn\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.869057 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:18 crc kubenswrapper[4822]: I1124 14:23:18.938415 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nvv4m"] Nov 24 14:23:18 crc kubenswrapper[4822]: W1124 14:23:18.945573 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39db4f3b_fa4e_41ad_92c9_5e682207c429.slice/crio-47e8af4e6adfc12cc75a51f34bc142c178e7f417e3f9835bc98984d334696aad WatchSource:0}: Error finding container 47e8af4e6adfc12cc75a51f34bc142c178e7f417e3f9835bc98984d334696aad: Status 404 returned error can't find the container with id 47e8af4e6adfc12cc75a51f34bc142c178e7f417e3f9835bc98984d334696aad Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.059112 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9mzfn"] Nov 24 14:23:19 crc kubenswrapper[4822]: W1124 14:23:19.062105 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5e9ae7c_8cf5_404a_8403_94c5005ade7e.slice/crio-c896ddde049b808c6d0f2bf4329657ea54a6782341b32409b2af46f0eefc7bab WatchSource:0}: Error finding container c896ddde049b808c6d0f2bf4329657ea54a6782341b32409b2af46f0eefc7bab: Status 404 returned error can't find the container with id c896ddde049b808c6d0f2bf4329657ea54a6782341b32409b2af46f0eefc7bab Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.402431 4822 generic.go:334] "Generic (PLEG): container finished" podID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerID="c0e417a46e8b3c6cf52e73785427c0e4b4f04cd29a2158880558de373bdf762f" exitCode=0 Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.402503 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mzfn" event={"ID":"d5e9ae7c-8cf5-404a-8403-94c5005ade7e","Type":"ContainerDied","Data":"c0e417a46e8b3c6cf52e73785427c0e4b4f04cd29a2158880558de373bdf762f"} Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.402537 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mzfn" event={"ID":"d5e9ae7c-8cf5-404a-8403-94c5005ade7e","Type":"ContainerStarted","Data":"c896ddde049b808c6d0f2bf4329657ea54a6782341b32409b2af46f0eefc7bab"} Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.407891 4822 generic.go:334] "Generic (PLEG): container finished" podID="e5fca0b7-7cab-4a35-a0ef-88a75d92707a" containerID="7dc0fcdcaa1494ab3ae99aad5144be6bf1244d2741805526a5828a518d241a8e" exitCode=0 Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.407978 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlfkr" event={"ID":"e5fca0b7-7cab-4a35-a0ef-88a75d92707a","Type":"ContainerDied","Data":"7dc0fcdcaa1494ab3ae99aad5144be6bf1244d2741805526a5828a518d241a8e"} Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.410088 4822 generic.go:334] "Generic (PLEG): container finished" podID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerID="93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0" exitCode=0 Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.410139 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28rbg" event={"ID":"5c76e331-f8e8-45e9-9afb-2add7327ebd1","Type":"ContainerDied","Data":"93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0"} Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.418282 4822 generic.go:334] "Generic (PLEG): container finished" podID="39db4f3b-fa4e-41ad-92c9-5e682207c429" containerID="3c3087b66d11de74dab09ab7695458c5fe70d2f47404482453f0ec374d4cdcf1" exitCode=0 Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.418333 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvv4m" event={"ID":"39db4f3b-fa4e-41ad-92c9-5e682207c429","Type":"ContainerDied","Data":"3c3087b66d11de74dab09ab7695458c5fe70d2f47404482453f0ec374d4cdcf1"} Nov 24 14:23:19 crc kubenswrapper[4822]: I1124 14:23:19.418361 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvv4m" event={"ID":"39db4f3b-fa4e-41ad-92c9-5e682207c429","Type":"ContainerStarted","Data":"47e8af4e6adfc12cc75a51f34bc142c178e7f417e3f9835bc98984d334696aad"} Nov 24 14:23:20 crc kubenswrapper[4822]: I1124 14:23:20.426666 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28rbg" event={"ID":"5c76e331-f8e8-45e9-9afb-2add7327ebd1","Type":"ContainerStarted","Data":"d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0"} Nov 24 14:23:20 crc kubenswrapper[4822]: I1124 14:23:20.429330 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvv4m" event={"ID":"39db4f3b-fa4e-41ad-92c9-5e682207c429","Type":"ContainerStarted","Data":"7afe4c918d0cb259639a56d1b4be118820399cf96f2ed303204ff808a4e32647"} Nov 24 14:23:20 crc kubenswrapper[4822]: I1124 14:23:20.431644 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mzfn" event={"ID":"d5e9ae7c-8cf5-404a-8403-94c5005ade7e","Type":"ContainerStarted","Data":"003768f6d9db7fa91bf6d2e91e54f0d8a74c109847687629c5027c3bcc35cc70"} Nov 24 14:23:20 crc kubenswrapper[4822]: I1124 14:23:20.436946 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlfkr" event={"ID":"e5fca0b7-7cab-4a35-a0ef-88a75d92707a","Type":"ContainerStarted","Data":"03ed119d92abe9f62615e9ccb9b961e33af87b32e47f4beb1596f540fbffe5d1"} Nov 24 14:23:20 crc kubenswrapper[4822]: I1124 14:23:20.450103 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-28rbg" podStartSLOduration=2.92525456 podStartE2EDuration="5.450086507s" podCreationTimestamp="2025-11-24 14:23:15 +0000 UTC" firstStartedPulling="2025-11-24 14:23:17.378012386 +0000 UTC m=+234.494652863" lastFinishedPulling="2025-11-24 14:23:19.902844323 +0000 UTC m=+237.019484810" observedRunningTime="2025-11-24 14:23:20.447314778 +0000 UTC m=+237.563955255" watchObservedRunningTime="2025-11-24 14:23:20.450086507 +0000 UTC m=+237.566726984" Nov 24 14:23:20 crc kubenswrapper[4822]: I1124 14:23:20.490514 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vlfkr" podStartSLOduration=2.017152262 podStartE2EDuration="4.490495601s" podCreationTimestamp="2025-11-24 14:23:16 +0000 UTC" firstStartedPulling="2025-11-24 14:23:17.383576524 +0000 UTC m=+234.500217021" lastFinishedPulling="2025-11-24 14:23:19.856919853 +0000 UTC m=+236.973560360" observedRunningTime="2025-11-24 14:23:20.488328402 +0000 UTC m=+237.604968899" watchObservedRunningTime="2025-11-24 14:23:20.490495601 +0000 UTC m=+237.607136078" Nov 24 14:23:21 crc kubenswrapper[4822]: I1124 14:23:21.446118 4822 generic.go:334] "Generic (PLEG): container finished" podID="39db4f3b-fa4e-41ad-92c9-5e682207c429" containerID="7afe4c918d0cb259639a56d1b4be118820399cf96f2ed303204ff808a4e32647" exitCode=0 Nov 24 14:23:21 crc kubenswrapper[4822]: I1124 14:23:21.446197 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvv4m" event={"ID":"39db4f3b-fa4e-41ad-92c9-5e682207c429","Type":"ContainerDied","Data":"7afe4c918d0cb259639a56d1b4be118820399cf96f2ed303204ff808a4e32647"} Nov 24 14:23:21 crc kubenswrapper[4822]: I1124 14:23:21.450689 4822 generic.go:334] "Generic (PLEG): container finished" podID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerID="003768f6d9db7fa91bf6d2e91e54f0d8a74c109847687629c5027c3bcc35cc70" exitCode=0 Nov 24 14:23:21 crc kubenswrapper[4822]: I1124 14:23:21.451390 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mzfn" event={"ID":"d5e9ae7c-8cf5-404a-8403-94c5005ade7e","Type":"ContainerDied","Data":"003768f6d9db7fa91bf6d2e91e54f0d8a74c109847687629c5027c3bcc35cc70"} Nov 24 14:23:23 crc kubenswrapper[4822]: I1124 14:23:23.462404 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvv4m" event={"ID":"39db4f3b-fa4e-41ad-92c9-5e682207c429","Type":"ContainerStarted","Data":"b031229abf0aca63305e30d14bf578c6116fa8b3ca59401fcd55899330a0b29c"} Nov 24 14:23:23 crc kubenswrapper[4822]: I1124 14:23:23.463744 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mzfn" event={"ID":"d5e9ae7c-8cf5-404a-8403-94c5005ade7e","Type":"ContainerStarted","Data":"822a597b520e7a18214ee4487b78524bac10e3725fb3a0d2795a6cf5f81a9eaf"} Nov 24 14:23:23 crc kubenswrapper[4822]: I1124 14:23:23.511998 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nvv4m" podStartSLOduration=3.080052038 podStartE2EDuration="5.511969622s" podCreationTimestamp="2025-11-24 14:23:18 +0000 UTC" firstStartedPulling="2025-11-24 14:23:19.419305059 +0000 UTC m=+236.535945536" lastFinishedPulling="2025-11-24 14:23:21.851222623 +0000 UTC m=+238.967863120" observedRunningTime="2025-11-24 14:23:23.487350334 +0000 UTC m=+240.603990831" watchObservedRunningTime="2025-11-24 14:23:23.511969622 +0000 UTC m=+240.628610099" Nov 24 14:23:23 crc kubenswrapper[4822]: I1124 14:23:23.512296 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9mzfn" podStartSLOduration=2.938977223 podStartE2EDuration="5.512290473s" podCreationTimestamp="2025-11-24 14:23:18 +0000 UTC" firstStartedPulling="2025-11-24 14:23:19.404725313 +0000 UTC m=+236.521365830" lastFinishedPulling="2025-11-24 14:23:21.978038593 +0000 UTC m=+239.094679080" observedRunningTime="2025-11-24 14:23:23.502866931 +0000 UTC m=+240.619507428" watchObservedRunningTime="2025-11-24 14:23:23.512290473 +0000 UTC m=+240.628930950" Nov 24 14:23:26 crc kubenswrapper[4822]: I1124 14:23:26.268516 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:26 crc kubenswrapper[4822]: I1124 14:23:26.269145 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:26 crc kubenswrapper[4822]: I1124 14:23:26.309663 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:26 crc kubenswrapper[4822]: I1124 14:23:26.469300 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:26 crc kubenswrapper[4822]: I1124 14:23:26.469368 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:26 crc kubenswrapper[4822]: I1124 14:23:26.525124 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-28rbg" Nov 24 14:23:26 crc kubenswrapper[4822]: I1124 14:23:26.528718 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:26 crc kubenswrapper[4822]: I1124 14:23:26.592965 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vlfkr" Nov 24 14:23:28 crc kubenswrapper[4822]: I1124 14:23:28.750104 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:28 crc kubenswrapper[4822]: I1124 14:23:28.750502 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:28 crc kubenswrapper[4822]: I1124 14:23:28.800828 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:28 crc kubenswrapper[4822]: I1124 14:23:28.870149 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:28 crc kubenswrapper[4822]: I1124 14:23:28.870225 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:28 crc kubenswrapper[4822]: I1124 14:23:28.932307 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:23:29 crc kubenswrapper[4822]: I1124 14:23:29.538791 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nvv4m" Nov 24 14:23:29 crc kubenswrapper[4822]: I1124 14:23:29.549907 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:24:23 crc kubenswrapper[4822]: I1124 14:24:23.489458 4822 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 24 14:25:11 crc kubenswrapper[4822]: I1124 14:25:11.578628 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:25:11 crc kubenswrapper[4822]: I1124 14:25:11.579246 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:25:41 crc kubenswrapper[4822]: I1124 14:25:41.578575 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:25:41 crc kubenswrapper[4822]: I1124 14:25:41.579279 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:26:11 crc kubenswrapper[4822]: I1124 14:26:11.577943 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:26:11 crc kubenswrapper[4822]: I1124 14:26:11.578584 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:26:11 crc kubenswrapper[4822]: I1124 14:26:11.578650 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:26:11 crc kubenswrapper[4822]: I1124 14:26:11.579310 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7115168ad50bfd88ed09dd9295ac9935a4ada79bd3272f63430ce467dc596e9b"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:26:11 crc kubenswrapper[4822]: I1124 14:26:11.579383 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://7115168ad50bfd88ed09dd9295ac9935a4ada79bd3272f63430ce467dc596e9b" gracePeriod=600 Nov 24 14:26:12 crc kubenswrapper[4822]: I1124 14:26:12.568305 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="7115168ad50bfd88ed09dd9295ac9935a4ada79bd3272f63430ce467dc596e9b" exitCode=0 Nov 24 14:26:12 crc kubenswrapper[4822]: I1124 14:26:12.568396 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"7115168ad50bfd88ed09dd9295ac9935a4ada79bd3272f63430ce467dc596e9b"} Nov 24 14:26:12 crc kubenswrapper[4822]: I1124 14:26:12.568865 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"c48afa8921812e0254372f9525778df4abed8a487a31452c711ba4cfd0103ae3"} Nov 24 14:26:12 crc kubenswrapper[4822]: I1124 14:26:12.568901 4822 scope.go:117] "RemoveContainer" containerID="2cb16cd81d175a3967132357472b7a8a9f25c65a4fedea6221fd1a9b2882e54f" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.175879 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nhngk"] Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.177652 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.194159 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nhngk"] Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.327553 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9a4cec7d-db87-47c2-be56-57368b7d9768-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.327637 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdlb6\" (UniqueName: \"kubernetes.io/projected/9a4cec7d-db87-47c2-be56-57368b7d9768-kube-api-access-bdlb6\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.327657 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9a4cec7d-db87-47c2-be56-57368b7d9768-registry-tls\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.327756 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9a4cec7d-db87-47c2-be56-57368b7d9768-registry-certificates\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.327821 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9a4cec7d-db87-47c2-be56-57368b7d9768-bound-sa-token\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.327863 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a4cec7d-db87-47c2-be56-57368b7d9768-trusted-ca\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.328061 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9a4cec7d-db87-47c2-be56-57368b7d9768-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.328148 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.356252 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.428777 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9a4cec7d-db87-47c2-be56-57368b7d9768-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.428874 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9a4cec7d-db87-47c2-be56-57368b7d9768-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.428912 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9a4cec7d-db87-47c2-be56-57368b7d9768-registry-tls\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.428931 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdlb6\" (UniqueName: \"kubernetes.io/projected/9a4cec7d-db87-47c2-be56-57368b7d9768-kube-api-access-bdlb6\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.428954 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9a4cec7d-db87-47c2-be56-57368b7d9768-registry-certificates\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.428982 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9a4cec7d-db87-47c2-be56-57368b7d9768-bound-sa-token\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.429006 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a4cec7d-db87-47c2-be56-57368b7d9768-trusted-ca\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.429505 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9a4cec7d-db87-47c2-be56-57368b7d9768-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.430098 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a4cec7d-db87-47c2-be56-57368b7d9768-trusted-ca\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.430248 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9a4cec7d-db87-47c2-be56-57368b7d9768-registry-certificates\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.436304 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9a4cec7d-db87-47c2-be56-57368b7d9768-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.440190 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9a4cec7d-db87-47c2-be56-57368b7d9768-registry-tls\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.445512 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdlb6\" (UniqueName: \"kubernetes.io/projected/9a4cec7d-db87-47c2-be56-57368b7d9768-kube-api-access-bdlb6\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.446160 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9a4cec7d-db87-47c2-be56-57368b7d9768-bound-sa-token\") pod \"image-registry-66df7c8f76-nhngk\" (UID: \"9a4cec7d-db87-47c2-be56-57368b7d9768\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.496463 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.672027 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nhngk"] Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.853508 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" event={"ID":"9a4cec7d-db87-47c2-be56-57368b7d9768","Type":"ContainerStarted","Data":"5710a04979016507802d938b99ecf55b3f5ff75af6e832b021f08b45922b5749"} Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.853561 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" event={"ID":"9a4cec7d-db87-47c2-be56-57368b7d9768","Type":"ContainerStarted","Data":"2ef62e98acde804ed06caa3b319c6e434afe2314e36dc6978033164b42052813"} Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.853638 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:27:49 crc kubenswrapper[4822]: I1124 14:27:49.877926 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" podStartSLOduration=0.877903873 podStartE2EDuration="877.903873ms" podCreationTimestamp="2025-11-24 14:27:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:27:49.876942223 +0000 UTC m=+506.993582740" watchObservedRunningTime="2025-11-24 14:27:49.877903873 +0000 UTC m=+506.994544350" Nov 24 14:28:09 crc kubenswrapper[4822]: I1124 14:28:09.505515 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-nhngk" Nov 24 14:28:09 crc kubenswrapper[4822]: I1124 14:28:09.584914 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2z5sq"] Nov 24 14:28:11 crc kubenswrapper[4822]: I1124 14:28:11.578628 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:28:11 crc kubenswrapper[4822]: I1124 14:28:11.578701 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:28:34 crc kubenswrapper[4822]: I1124 14:28:34.633856 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" podUID="d5224ecc-afd0-4226-acd8-cb2c4197639c" containerName="registry" containerID="cri-o://b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660" gracePeriod=30 Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.026555 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.084605 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-trusted-ca\") pod \"d5224ecc-afd0-4226-acd8-cb2c4197639c\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.085032 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5224ecc-afd0-4226-acd8-cb2c4197639c-ca-trust-extracted\") pod \"d5224ecc-afd0-4226-acd8-cb2c4197639c\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.085283 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d5224ecc-afd0-4226-acd8-cb2c4197639c\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.085351 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhvwl\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-kube-api-access-dhvwl\") pod \"d5224ecc-afd0-4226-acd8-cb2c4197639c\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.085436 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-tls\") pod \"d5224ecc-afd0-4226-acd8-cb2c4197639c\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.085484 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-bound-sa-token\") pod \"d5224ecc-afd0-4226-acd8-cb2c4197639c\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.085549 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5224ecc-afd0-4226-acd8-cb2c4197639c-installation-pull-secrets\") pod \"d5224ecc-afd0-4226-acd8-cb2c4197639c\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.085616 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-certificates\") pod \"d5224ecc-afd0-4226-acd8-cb2c4197639c\" (UID: \"d5224ecc-afd0-4226-acd8-cb2c4197639c\") " Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.086966 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d5224ecc-afd0-4226-acd8-cb2c4197639c" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.088499 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.089325 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d5224ecc-afd0-4226-acd8-cb2c4197639c" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.095947 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d5224ecc-afd0-4226-acd8-cb2c4197639c" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.107938 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d5224ecc-afd0-4226-acd8-cb2c4197639c" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.109867 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d5224ecc-afd0-4226-acd8-cb2c4197639c" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.117387 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5224ecc-afd0-4226-acd8-cb2c4197639c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d5224ecc-afd0-4226-acd8-cb2c4197639c" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.117591 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-kube-api-access-dhvwl" (OuterVolumeSpecName: "kube-api-access-dhvwl") pod "d5224ecc-afd0-4226-acd8-cb2c4197639c" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c"). InnerVolumeSpecName "kube-api-access-dhvwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.118056 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5224ecc-afd0-4226-acd8-cb2c4197639c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d5224ecc-afd0-4226-acd8-cb2c4197639c" (UID: "d5224ecc-afd0-4226-acd8-cb2c4197639c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.141813 4822 generic.go:334] "Generic (PLEG): container finished" podID="d5224ecc-afd0-4226-acd8-cb2c4197639c" containerID="b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660" exitCode=0 Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.141907 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" event={"ID":"d5224ecc-afd0-4226-acd8-cb2c4197639c","Type":"ContainerDied","Data":"b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660"} Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.141948 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" event={"ID":"d5224ecc-afd0-4226-acd8-cb2c4197639c","Type":"ContainerDied","Data":"3dab6393b3f64b698f899af77fc0e727f25b64ac63c9e0b5816dd18e3e8546fb"} Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.141908 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2z5sq" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.141973 4822 scope.go:117] "RemoveContainer" containerID="b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.167306 4822 scope.go:117] "RemoveContainer" containerID="b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660" Nov 24 14:28:35 crc kubenswrapper[4822]: E1124 14:28:35.168198 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660\": container with ID starting with b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660 not found: ID does not exist" containerID="b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.168295 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660"} err="failed to get container status \"b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660\": rpc error: code = NotFound desc = could not find container \"b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660\": container with ID starting with b060f5946fbf18202e738ba72e11ab97f597122023620e713f63125114446660 not found: ID does not exist" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.188674 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2z5sq"] Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.189441 4822 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.189498 4822 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.189523 4822 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5224ecc-afd0-4226-acd8-cb2c4197639c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.189548 4822 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5224ecc-afd0-4226-acd8-cb2c4197639c-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.189563 4822 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5224ecc-afd0-4226-acd8-cb2c4197639c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.189579 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhvwl\" (UniqueName: \"kubernetes.io/projected/d5224ecc-afd0-4226-acd8-cb2c4197639c-kube-api-access-dhvwl\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.192001 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2z5sq"] Nov 24 14:28:35 crc kubenswrapper[4822]: I1124 14:28:35.713316 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5224ecc-afd0-4226-acd8-cb2c4197639c" path="/var/lib/kubelet/pods/d5224ecc-afd0-4226-acd8-cb2c4197639c/volumes" Nov 24 14:28:41 crc kubenswrapper[4822]: I1124 14:28:41.577812 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:28:41 crc kubenswrapper[4822]: I1124 14:28:41.578467 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:28:43 crc kubenswrapper[4822]: I1124 14:28:43.812754 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l"] Nov 24 14:28:43 crc kubenswrapper[4822]: E1124 14:28:43.813100 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5224ecc-afd0-4226-acd8-cb2c4197639c" containerName="registry" Nov 24 14:28:43 crc kubenswrapper[4822]: I1124 14:28:43.813123 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5224ecc-afd0-4226-acd8-cb2c4197639c" containerName="registry" Nov 24 14:28:43 crc kubenswrapper[4822]: I1124 14:28:43.813319 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5224ecc-afd0-4226-acd8-cb2c4197639c" containerName="registry" Nov 24 14:28:43 crc kubenswrapper[4822]: I1124 14:28:43.814583 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:43 crc kubenswrapper[4822]: I1124 14:28:43.817509 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 14:28:43 crc kubenswrapper[4822]: I1124 14:28:43.820506 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l"] Nov 24 14:28:43 crc kubenswrapper[4822]: I1124 14:28:43.903604 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:43 crc kubenswrapper[4822]: I1124 14:28:43.903702 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:43 crc kubenswrapper[4822]: I1124 14:28:43.903755 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m86rm\" (UniqueName: \"kubernetes.io/projected/a69ba74b-663d-4777-b50e-6988f9a9a894-kube-api-access-m86rm\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:44 crc kubenswrapper[4822]: I1124 14:28:44.004283 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:44 crc kubenswrapper[4822]: I1124 14:28:44.004331 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m86rm\" (UniqueName: \"kubernetes.io/projected/a69ba74b-663d-4777-b50e-6988f9a9a894-kube-api-access-m86rm\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:44 crc kubenswrapper[4822]: I1124 14:28:44.004380 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:44 crc kubenswrapper[4822]: I1124 14:28:44.004998 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:44 crc kubenswrapper[4822]: I1124 14:28:44.005004 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:44 crc kubenswrapper[4822]: I1124 14:28:44.036961 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m86rm\" (UniqueName: \"kubernetes.io/projected/a69ba74b-663d-4777-b50e-6988f9a9a894-kube-api-access-m86rm\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:44 crc kubenswrapper[4822]: I1124 14:28:44.141607 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:44 crc kubenswrapper[4822]: I1124 14:28:44.403492 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l"] Nov 24 14:28:45 crc kubenswrapper[4822]: I1124 14:28:45.220402 4822 generic.go:334] "Generic (PLEG): container finished" podID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerID="1a94077167cd408a5f9bbca80572f2b03a94dc447ba7be419b973c7eb4576076" exitCode=0 Nov 24 14:28:45 crc kubenswrapper[4822]: I1124 14:28:45.220519 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" event={"ID":"a69ba74b-663d-4777-b50e-6988f9a9a894","Type":"ContainerDied","Data":"1a94077167cd408a5f9bbca80572f2b03a94dc447ba7be419b973c7eb4576076"} Nov 24 14:28:45 crc kubenswrapper[4822]: I1124 14:28:45.220751 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" event={"ID":"a69ba74b-663d-4777-b50e-6988f9a9a894","Type":"ContainerStarted","Data":"42676a603422b7e1f1f9db93e7343ecf8151f5188c62dc64165a2ab80257ad87"} Nov 24 14:28:45 crc kubenswrapper[4822]: I1124 14:28:45.222714 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:28:47 crc kubenswrapper[4822]: I1124 14:28:47.237562 4822 generic.go:334] "Generic (PLEG): container finished" podID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerID="43e6355d54210095ef0fec0db414210b6fab53df845654ddc26aae65a11e6aed" exitCode=0 Nov 24 14:28:47 crc kubenswrapper[4822]: I1124 14:28:47.237719 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" event={"ID":"a69ba74b-663d-4777-b50e-6988f9a9a894","Type":"ContainerDied","Data":"43e6355d54210095ef0fec0db414210b6fab53df845654ddc26aae65a11e6aed"} Nov 24 14:28:48 crc kubenswrapper[4822]: I1124 14:28:48.247094 4822 generic.go:334] "Generic (PLEG): container finished" podID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerID="6222d51283d534e6d26a2de1de217110ba0ad5672a0ad09f08f33304e07573ff" exitCode=0 Nov 24 14:28:48 crc kubenswrapper[4822]: I1124 14:28:48.247194 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" event={"ID":"a69ba74b-663d-4777-b50e-6988f9a9a894","Type":"ContainerDied","Data":"6222d51283d534e6d26a2de1de217110ba0ad5672a0ad09f08f33304e07573ff"} Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.550809 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.592183 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m86rm\" (UniqueName: \"kubernetes.io/projected/a69ba74b-663d-4777-b50e-6988f9a9a894-kube-api-access-m86rm\") pod \"a69ba74b-663d-4777-b50e-6988f9a9a894\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.592344 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-bundle\") pod \"a69ba74b-663d-4777-b50e-6988f9a9a894\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.592406 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-util\") pod \"a69ba74b-663d-4777-b50e-6988f9a9a894\" (UID: \"a69ba74b-663d-4777-b50e-6988f9a9a894\") " Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.595710 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-bundle" (OuterVolumeSpecName: "bundle") pod "a69ba74b-663d-4777-b50e-6988f9a9a894" (UID: "a69ba74b-663d-4777-b50e-6988f9a9a894"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.603311 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a69ba74b-663d-4777-b50e-6988f9a9a894-kube-api-access-m86rm" (OuterVolumeSpecName: "kube-api-access-m86rm") pod "a69ba74b-663d-4777-b50e-6988f9a9a894" (UID: "a69ba74b-663d-4777-b50e-6988f9a9a894"). InnerVolumeSpecName "kube-api-access-m86rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.615747 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-util" (OuterVolumeSpecName: "util") pod "a69ba74b-663d-4777-b50e-6988f9a9a894" (UID: "a69ba74b-663d-4777-b50e-6988f9a9a894"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.693894 4822 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.693945 4822 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a69ba74b-663d-4777-b50e-6988f9a9a894-util\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:49 crc kubenswrapper[4822]: I1124 14:28:49.693963 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m86rm\" (UniqueName: \"kubernetes.io/projected/a69ba74b-663d-4777-b50e-6988f9a9a894-kube-api-access-m86rm\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:50 crc kubenswrapper[4822]: I1124 14:28:50.264019 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" event={"ID":"a69ba74b-663d-4777-b50e-6988f9a9a894","Type":"ContainerDied","Data":"42676a603422b7e1f1f9db93e7343ecf8151f5188c62dc64165a2ab80257ad87"} Nov 24 14:28:50 crc kubenswrapper[4822]: I1124 14:28:50.264084 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42676a603422b7e1f1f9db93e7343ecf8151f5188c62dc64165a2ab80257ad87" Nov 24 14:28:50 crc kubenswrapper[4822]: I1124 14:28:50.264112 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.054902 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-swpcw"] Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.055913 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovn-controller" containerID="cri-o://63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766" gracePeriod=30 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.055984 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="nbdb" containerID="cri-o://9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767" gracePeriod=30 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.056036 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a" gracePeriod=30 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.056067 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="northd" containerID="cri-o://046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8" gracePeriod=30 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.056128 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="sbdb" containerID="cri-o://38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec" gracePeriod=30 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.056171 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kube-rbac-proxy-node" containerID="cri-o://ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d" gracePeriod=30 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.056489 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovn-acl-logging" containerID="cri-o://cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57" gracePeriod=30 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.111056 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" containerID="cri-o://7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1" gracePeriod=30 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.296499 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/2.log" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.297295 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/1.log" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.297345 4822 generic.go:334] "Generic (PLEG): container finished" podID="92c230aa-aed2-4b07-82bd-98b1d2ffa456" containerID="f528354f90b6adf0611b43f5ddea194a67abb6f33c1715df7b8dc36dab6a88f9" exitCode=2 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.297402 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-d6rbt" event={"ID":"92c230aa-aed2-4b07-82bd-98b1d2ffa456","Type":"ContainerDied","Data":"f528354f90b6adf0611b43f5ddea194a67abb6f33c1715df7b8dc36dab6a88f9"} Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.297450 4822 scope.go:117] "RemoveContainer" containerID="6234753e9993e764870511b16c5d452771c0b7ea844499a737168536f11895f5" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.298108 4822 scope.go:117] "RemoveContainer" containerID="f528354f90b6adf0611b43f5ddea194a67abb6f33c1715df7b8dc36dab6a88f9" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.298382 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-d6rbt_openshift-multus(92c230aa-aed2-4b07-82bd-98b1d2ffa456)\"" pod="openshift-multus/multus-d6rbt" podUID="92c230aa-aed2-4b07-82bd-98b1d2ffa456" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.301924 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovnkube-controller/3.log" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.304681 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovn-acl-logging/0.log" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305244 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovn-controller/0.log" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305647 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1" exitCode=0 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305680 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a" exitCode=0 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305692 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d" exitCode=0 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305702 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57" exitCode=143 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305712 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766" exitCode=143 Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305736 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1"} Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305771 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a"} Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305788 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d"} Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305800 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57"} Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.305812 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766"} Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.343315 4822 scope.go:117] "RemoveContainer" containerID="3efd51ae7ad29f2c2b7c7b020482c678f390abe83cb7e657996c7b547b6f53aa" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.471252 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovn-acl-logging/0.log" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.472097 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovn-controller/0.log" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.472624 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558433 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7m9hl"] Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558679 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="sbdb" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558694 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="sbdb" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558708 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558717 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558726 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="nbdb" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558734 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="nbdb" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558743 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558751 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558761 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kubecfg-setup" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558768 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kubecfg-setup" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558777 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558784 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558793 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovn-acl-logging" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558800 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovn-acl-logging" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558815 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovn-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558822 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovn-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558833 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kube-rbac-proxy-node" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558840 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kube-rbac-proxy-node" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558851 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558860 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558875 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="northd" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558883 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="northd" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558893 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerName="extract" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558901 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerName="extract" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558909 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerName="pull" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558917 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerName="pull" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.558929 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerName="util" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.558938 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerName="util" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559081 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kube-rbac-proxy-node" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559095 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="sbdb" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559106 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559116 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559126 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559134 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69ba74b-663d-4777-b50e-6988f9a9a894" containerName="extract" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559146 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovn-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559158 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="northd" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559166 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="nbdb" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559178 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovn-acl-logging" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559187 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.559370 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559381 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559498 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559508 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: E1124 14:28:55.559643 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.559653 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerName="ovnkube-controller" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.562438 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569286 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-openvswitch\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569343 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-node-log\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569379 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-systemd-units\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569412 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhtnb\" (UniqueName: \"kubernetes.io/projected/13d10aa8-fab6-40df-90c8-0784ebac8e0f-kube-api-access-nhtnb\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569445 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-config\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569478 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-env-overrides\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569545 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-systemd\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569577 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovn-node-metrics-cert\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569614 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-etc-openvswitch\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569637 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-slash\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569669 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569693 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-netns\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569922 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-node-log" (OuterVolumeSpecName: "node-log") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569973 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569994 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.570496 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.570566 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.570608 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-slash" (OuterVolumeSpecName: "host-slash") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.570645 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.570758 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.569721 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-ovn-kubernetes\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571354 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-var-lib-openvswitch\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571292 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571428 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-ovn\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571355 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571470 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-kubelet\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571512 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-bin\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571527 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571596 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571628 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-netd\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571666 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-script-lib\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571662 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571699 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-log-socket\") pod \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\" (UID: \"13d10aa8-fab6-40df-90c8-0784ebac8e0f\") " Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571730 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-log-socket" (OuterVolumeSpecName: "log-socket") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.571763 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572240 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572290 4822 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572322 4822 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572343 4822 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572362 4822 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572381 4822 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572398 4822 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572416 4822 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572433 4822 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572452 4822 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572472 4822 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572491 4822 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572511 4822 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572529 4822 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572546 4822 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572563 4822 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.572606 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.574771 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13d10aa8-fab6-40df-90c8-0784ebac8e0f-kube-api-access-nhtnb" (OuterVolumeSpecName: "kube-api-access-nhtnb") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "kube-api-access-nhtnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.578405 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.587669 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "13d10aa8-fab6-40df-90c8-0784ebac8e0f" (UID: "13d10aa8-fab6-40df-90c8-0784ebac8e0f"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.673188 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-kubelet\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.673556 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-run-netns\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.673679 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.673781 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d71763fd-618d-4c6c-b062-52fbffcb68d4-ovnkube-config\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.673891 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-slash\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.673976 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-run-systemd\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.674071 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-var-lib-openvswitch\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.674170 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv4ww\" (UniqueName: \"kubernetes.io/projected/d71763fd-618d-4c6c-b062-52fbffcb68d4-kube-api-access-sv4ww\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.674329 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-cni-netd\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.674432 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-cni-bin\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.674562 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d71763fd-618d-4c6c-b062-52fbffcb68d4-env-overrides\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.674693 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-run-openvswitch\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.674803 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-etc-openvswitch\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.674909 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d71763fd-618d-4c6c-b062-52fbffcb68d4-ovn-node-metrics-cert\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675012 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-run-ovn-kubernetes\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675110 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-systemd-units\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675231 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-node-log\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675326 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-log-socket\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675426 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d71763fd-618d-4c6c-b062-52fbffcb68d4-ovnkube-script-lib\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675523 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-run-ovn\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675645 4822 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675729 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhtnb\" (UniqueName: \"kubernetes.io/projected/13d10aa8-fab6-40df-90c8-0784ebac8e0f-kube-api-access-nhtnb\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675807 4822 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675879 4822 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13d10aa8-fab6-40df-90c8-0784ebac8e0f-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.675950 4822 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13d10aa8-fab6-40df-90c8-0784ebac8e0f-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.776702 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-etc-openvswitch\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.776999 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d71763fd-618d-4c6c-b062-52fbffcb68d4-ovn-node-metrics-cert\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.777601 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-run-ovn-kubernetes\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.777689 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-systemd-units\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.777790 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-node-log\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.777869 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-log-socket\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.777973 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-log-socket\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.777926 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-node-log\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.776838 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-etc-openvswitch\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.777698 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-run-ovn-kubernetes\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.777999 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d71763fd-618d-4c6c-b062-52fbffcb68d4-ovnkube-script-lib\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778293 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-run-ovn\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778410 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-kubelet\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778467 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-run-netns\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778518 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778573 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d71763fd-618d-4c6c-b062-52fbffcb68d4-ovnkube-config\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778669 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-run-systemd\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778725 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-slash\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778759 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-var-lib-openvswitch\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778801 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv4ww\" (UniqueName: \"kubernetes.io/projected/d71763fd-618d-4c6c-b062-52fbffcb68d4-kube-api-access-sv4ww\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778901 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-cni-netd\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.778958 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-cni-bin\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.779040 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d71763fd-618d-4c6c-b062-52fbffcb68d4-env-overrides\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.779103 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-run-openvswitch\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.779293 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-run-openvswitch\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.777852 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-systemd-units\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.779426 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-run-ovn\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.779497 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-kubelet\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.779562 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-run-netns\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.779632 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.779651 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d71763fd-618d-4c6c-b062-52fbffcb68d4-ovnkube-script-lib\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.779764 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-var-lib-openvswitch\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.780248 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d71763fd-618d-4c6c-b062-52fbffcb68d4-ovnkube-config\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.780295 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-cni-netd\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.780345 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-cni-bin\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.780470 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-run-systemd\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.780541 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d71763fd-618d-4c6c-b062-52fbffcb68d4-host-slash\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.781278 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d71763fd-618d-4c6c-b062-52fbffcb68d4-env-overrides\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.781747 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d71763fd-618d-4c6c-b062-52fbffcb68d4-ovn-node-metrics-cert\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.811374 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv4ww\" (UniqueName: \"kubernetes.io/projected/d71763fd-618d-4c6c-b062-52fbffcb68d4-kube-api-access-sv4ww\") pod \"ovnkube-node-7m9hl\" (UID: \"d71763fd-618d-4c6c-b062-52fbffcb68d4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:55 crc kubenswrapper[4822]: I1124 14:28:55.878108 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.311325 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/2.log" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.313320 4822 generic.go:334] "Generic (PLEG): container finished" podID="d71763fd-618d-4c6c-b062-52fbffcb68d4" containerID="e5260c6931d629fa0a89757e9225cddab452c784f677798c106aa52a56282c7b" exitCode=0 Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.313450 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerDied","Data":"e5260c6931d629fa0a89757e9225cddab452c784f677798c106aa52a56282c7b"} Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.313521 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerStarted","Data":"8242f012bc955245c75f6af2d6fae2f2e1141ff7acf10ae89383418f4de88310"} Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.319269 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovn-acl-logging/0.log" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.319669 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-swpcw_13d10aa8-fab6-40df-90c8-0784ebac8e0f/ovn-controller/0.log" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.323587 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec" exitCode=0 Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.323612 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767" exitCode=0 Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.323622 4822 generic.go:334] "Generic (PLEG): container finished" podID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" containerID="046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8" exitCode=0 Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.323647 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec"} Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.323668 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767"} Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.323679 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8"} Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.323689 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" event={"ID":"13d10aa8-fab6-40df-90c8-0784ebac8e0f","Type":"ContainerDied","Data":"c26d06b1492bc2eb2fff84b5cc56e70eeb295186b1a80d88755064083c67754a"} Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.323705 4822 scope.go:117] "RemoveContainer" containerID="7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.323702 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-swpcw" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.340020 4822 scope.go:117] "RemoveContainer" containerID="38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.374918 4822 scope.go:117] "RemoveContainer" containerID="9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.391191 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-swpcw"] Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.393963 4822 scope.go:117] "RemoveContainer" containerID="046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.398956 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-swpcw"] Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.411470 4822 scope.go:117] "RemoveContainer" containerID="52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.425195 4822 scope.go:117] "RemoveContainer" containerID="ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.439404 4822 scope.go:117] "RemoveContainer" containerID="cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.459263 4822 scope.go:117] "RemoveContainer" containerID="63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.504499 4822 scope.go:117] "RemoveContainer" containerID="f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.541963 4822 scope.go:117] "RemoveContainer" containerID="7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1" Nov 24 14:28:56 crc kubenswrapper[4822]: E1124 14:28:56.542800 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1\": container with ID starting with 7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1 not found: ID does not exist" containerID="7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.542846 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1"} err="failed to get container status \"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1\": rpc error: code = NotFound desc = could not find container \"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1\": container with ID starting with 7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.542876 4822 scope.go:117] "RemoveContainer" containerID="38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec" Nov 24 14:28:56 crc kubenswrapper[4822]: E1124 14:28:56.543914 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\": container with ID starting with 38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec not found: ID does not exist" containerID="38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.543960 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec"} err="failed to get container status \"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\": rpc error: code = NotFound desc = could not find container \"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\": container with ID starting with 38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.543983 4822 scope.go:117] "RemoveContainer" containerID="9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767" Nov 24 14:28:56 crc kubenswrapper[4822]: E1124 14:28:56.544581 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\": container with ID starting with 9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767 not found: ID does not exist" containerID="9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.544635 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767"} err="failed to get container status \"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\": rpc error: code = NotFound desc = could not find container \"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\": container with ID starting with 9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.544663 4822 scope.go:117] "RemoveContainer" containerID="046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8" Nov 24 14:28:56 crc kubenswrapper[4822]: E1124 14:28:56.544992 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\": container with ID starting with 046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8 not found: ID does not exist" containerID="046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.545023 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8"} err="failed to get container status \"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\": rpc error: code = NotFound desc = could not find container \"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\": container with ID starting with 046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.545042 4822 scope.go:117] "RemoveContainer" containerID="52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a" Nov 24 14:28:56 crc kubenswrapper[4822]: E1124 14:28:56.545286 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\": container with ID starting with 52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a not found: ID does not exist" containerID="52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.545311 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a"} err="failed to get container status \"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\": rpc error: code = NotFound desc = could not find container \"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\": container with ID starting with 52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.545329 4822 scope.go:117] "RemoveContainer" containerID="ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d" Nov 24 14:28:56 crc kubenswrapper[4822]: E1124 14:28:56.545797 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\": container with ID starting with ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d not found: ID does not exist" containerID="ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.545816 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d"} err="failed to get container status \"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\": rpc error: code = NotFound desc = could not find container \"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\": container with ID starting with ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.545829 4822 scope.go:117] "RemoveContainer" containerID="cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57" Nov 24 14:28:56 crc kubenswrapper[4822]: E1124 14:28:56.546059 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\": container with ID starting with cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57 not found: ID does not exist" containerID="cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.546077 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57"} err="failed to get container status \"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\": rpc error: code = NotFound desc = could not find container \"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\": container with ID starting with cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.546097 4822 scope.go:117] "RemoveContainer" containerID="63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766" Nov 24 14:28:56 crc kubenswrapper[4822]: E1124 14:28:56.546345 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\": container with ID starting with 63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766 not found: ID does not exist" containerID="63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.546364 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766"} err="failed to get container status \"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\": rpc error: code = NotFound desc = could not find container \"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\": container with ID starting with 63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.546377 4822 scope.go:117] "RemoveContainer" containerID="f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7" Nov 24 14:28:56 crc kubenswrapper[4822]: E1124 14:28:56.546580 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\": container with ID starting with f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7 not found: ID does not exist" containerID="f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.546611 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7"} err="failed to get container status \"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\": rpc error: code = NotFound desc = could not find container \"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\": container with ID starting with f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.546626 4822 scope.go:117] "RemoveContainer" containerID="7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.546801 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1"} err="failed to get container status \"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1\": rpc error: code = NotFound desc = could not find container \"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1\": container with ID starting with 7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.546818 4822 scope.go:117] "RemoveContainer" containerID="38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547033 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec"} err="failed to get container status \"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\": rpc error: code = NotFound desc = could not find container \"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\": container with ID starting with 38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547053 4822 scope.go:117] "RemoveContainer" containerID="9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547258 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767"} err="failed to get container status \"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\": rpc error: code = NotFound desc = could not find container \"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\": container with ID starting with 9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547276 4822 scope.go:117] "RemoveContainer" containerID="046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547480 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8"} err="failed to get container status \"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\": rpc error: code = NotFound desc = could not find container \"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\": container with ID starting with 046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547497 4822 scope.go:117] "RemoveContainer" containerID="52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547714 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a"} err="failed to get container status \"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\": rpc error: code = NotFound desc = could not find container \"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\": container with ID starting with 52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547733 4822 scope.go:117] "RemoveContainer" containerID="ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547931 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d"} err="failed to get container status \"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\": rpc error: code = NotFound desc = could not find container \"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\": container with ID starting with ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.547949 4822 scope.go:117] "RemoveContainer" containerID="cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.548130 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57"} err="failed to get container status \"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\": rpc error: code = NotFound desc = could not find container \"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\": container with ID starting with cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.548151 4822 scope.go:117] "RemoveContainer" containerID="63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.548345 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766"} err="failed to get container status \"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\": rpc error: code = NotFound desc = could not find container \"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\": container with ID starting with 63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.548363 4822 scope.go:117] "RemoveContainer" containerID="f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.548794 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7"} err="failed to get container status \"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\": rpc error: code = NotFound desc = could not find container \"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\": container with ID starting with f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.548813 4822 scope.go:117] "RemoveContainer" containerID="7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.548982 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1"} err="failed to get container status \"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1\": rpc error: code = NotFound desc = could not find container \"7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1\": container with ID starting with 7e5b750e5d84184828e12825bd35c9858dd08359033b898331c101cf2038b1d1 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.548997 4822 scope.go:117] "RemoveContainer" containerID="38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549155 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec"} err="failed to get container status \"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\": rpc error: code = NotFound desc = could not find container \"38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec\": container with ID starting with 38968be2e1b19db13e792cba8ca43e8ff4f168a3fb7d26ee45640fbb057c9dec not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549177 4822 scope.go:117] "RemoveContainer" containerID="9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549348 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767"} err="failed to get container status \"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\": rpc error: code = NotFound desc = could not find container \"9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767\": container with ID starting with 9c70f2354d9bde2e708abcc02fcb6de1684785bbcd097a3609d4e13628b9e767 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549363 4822 scope.go:117] "RemoveContainer" containerID="046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549535 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8"} err="failed to get container status \"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\": rpc error: code = NotFound desc = could not find container \"046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8\": container with ID starting with 046e24b77be96c2db037959c4b1dc3298c0e0ae3112aa1d40a7c45bd8b4bf2e8 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549552 4822 scope.go:117] "RemoveContainer" containerID="52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549763 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a"} err="failed to get container status \"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\": rpc error: code = NotFound desc = could not find container \"52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a\": container with ID starting with 52573811d849fab04d6553e2e5583537862e4499db8249bc82bf1ae7cefb583a not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549781 4822 scope.go:117] "RemoveContainer" containerID="ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549976 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d"} err="failed to get container status \"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\": rpc error: code = NotFound desc = could not find container \"ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d\": container with ID starting with ac1ba9c794dead4f80b825ebc3c4e58b25c1eeba35c5c62ef488d84f4b43c34d not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.549996 4822 scope.go:117] "RemoveContainer" containerID="cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.550243 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57"} err="failed to get container status \"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\": rpc error: code = NotFound desc = could not find container \"cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57\": container with ID starting with cda07ab4518e06584efde01358d27dd970107dcd0b26568eacbab0ac90beff57 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.550263 4822 scope.go:117] "RemoveContainer" containerID="63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.550570 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766"} err="failed to get container status \"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\": rpc error: code = NotFound desc = could not find container \"63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766\": container with ID starting with 63afe498eaa69810b094df43c2b6f65e69ace7385c977477b39e7a84bb690766 not found: ID does not exist" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.550591 4822 scope.go:117] "RemoveContainer" containerID="f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7" Nov 24 14:28:56 crc kubenswrapper[4822]: I1124 14:28:56.554314 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7"} err="failed to get container status \"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\": rpc error: code = NotFound desc = could not find container \"f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7\": container with ID starting with f22fa0f878eb409bb7bdf9ffc13d27f45e2a49958e230fba90328c4139d4b8a7 not found: ID does not exist" Nov 24 14:28:57 crc kubenswrapper[4822]: I1124 14:28:57.333019 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerStarted","Data":"9536b809fe9ee078d3299d33261e7e4f863685f269cb35e31b89dc163c58578e"} Nov 24 14:28:57 crc kubenswrapper[4822]: I1124 14:28:57.333425 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerStarted","Data":"e577b441248e8cca8c6ce283274a6407f4b90749839e211bdcdfb996fd2c3e2a"} Nov 24 14:28:57 crc kubenswrapper[4822]: I1124 14:28:57.333441 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerStarted","Data":"522f2a74fe6b9775b34cd6223e5aaa0c9ff70136789d17024b5928dee59fca5e"} Nov 24 14:28:57 crc kubenswrapper[4822]: I1124 14:28:57.333460 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerStarted","Data":"7295841d57d85aef0c22b040f024c46a4f687ca363dd81f1452153ec652d8dbd"} Nov 24 14:28:57 crc kubenswrapper[4822]: I1124 14:28:57.333474 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerStarted","Data":"595838cb7c72b5381b251914a1d288f18a3e3c18b10079891eb7c22b8ea3294a"} Nov 24 14:28:57 crc kubenswrapper[4822]: I1124 14:28:57.333486 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerStarted","Data":"d3b348d0221545df1f70754b7c7e47e56421e15409951628eed3f112a273c1bc"} Nov 24 14:28:57 crc kubenswrapper[4822]: I1124 14:28:57.712013 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13d10aa8-fab6-40df-90c8-0784ebac8e0f" path="/var/lib/kubelet/pods/13d10aa8-fab6-40df-90c8-0784ebac8e0f/volumes" Nov 24 14:28:59 crc kubenswrapper[4822]: I1124 14:28:59.394558 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerStarted","Data":"d0e35104e1fbcba714083fdb56291d54d4fc5df76d8cbbf16288c2d8aee5d0c2"} Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.596351 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t"] Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.597449 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.601112 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.602220 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.605250 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-9ktzq" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.644910 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm"] Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.645951 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.648629 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.649013 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-g8mcw" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.652454 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7sql\" (UniqueName: \"kubernetes.io/projected/387c7641-9e5d-4ea0-a86f-c890d68d2903-kube-api-access-m7sql\") pod \"obo-prometheus-operator-668cf9dfbb-ql56t\" (UID: \"387c7641-9e5d-4ea0-a86f-c890d68d2903\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.669002 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd"] Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.669785 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.753432 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c463043b-1fc8-49c3-80a7-ac3469c1504a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm\" (UID: \"c463043b-1fc8-49c3-80a7-ac3469c1504a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.753490 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c463043b-1fc8-49c3-80a7-ac3469c1504a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm\" (UID: \"c463043b-1fc8-49c3-80a7-ac3469c1504a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.753508 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bdf59941-e2f3-4ccf-a700-992e2d3871af-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd\" (UID: \"bdf59941-e2f3-4ccf-a700-992e2d3871af\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.753534 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7sql\" (UniqueName: \"kubernetes.io/projected/387c7641-9e5d-4ea0-a86f-c890d68d2903-kube-api-access-m7sql\") pod \"obo-prometheus-operator-668cf9dfbb-ql56t\" (UID: \"387c7641-9e5d-4ea0-a86f-c890d68d2903\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.753581 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bdf59941-e2f3-4ccf-a700-992e2d3871af-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd\" (UID: \"bdf59941-e2f3-4ccf-a700-992e2d3871af\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.774022 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-sdtbh"] Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.774708 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.779559 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.779628 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-zpgvx" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.801070 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7sql\" (UniqueName: \"kubernetes.io/projected/387c7641-9e5d-4ea0-a86f-c890d68d2903-kube-api-access-m7sql\") pod \"obo-prometheus-operator-668cf9dfbb-ql56t\" (UID: \"387c7641-9e5d-4ea0-a86f-c890d68d2903\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.856822 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c463043b-1fc8-49c3-80a7-ac3469c1504a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm\" (UID: \"c463043b-1fc8-49c3-80a7-ac3469c1504a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.857434 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz8gk\" (UniqueName: \"kubernetes.io/projected/991aab37-0cca-4f05-a7ff-ff64d2cac3d4-kube-api-access-tz8gk\") pod \"observability-operator-d8bb48f5d-sdtbh\" (UID: \"991aab37-0cca-4f05-a7ff-ff64d2cac3d4\") " pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.857475 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c463043b-1fc8-49c3-80a7-ac3469c1504a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm\" (UID: \"c463043b-1fc8-49c3-80a7-ac3469c1504a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.857518 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bdf59941-e2f3-4ccf-a700-992e2d3871af-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd\" (UID: \"bdf59941-e2f3-4ccf-a700-992e2d3871af\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.857572 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bdf59941-e2f3-4ccf-a700-992e2d3871af-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd\" (UID: \"bdf59941-e2f3-4ccf-a700-992e2d3871af\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.857608 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/991aab37-0cca-4f05-a7ff-ff64d2cac3d4-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-sdtbh\" (UID: \"991aab37-0cca-4f05-a7ff-ff64d2cac3d4\") " pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.866854 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bdf59941-e2f3-4ccf-a700-992e2d3871af-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd\" (UID: \"bdf59941-e2f3-4ccf-a700-992e2d3871af\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.867778 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c463043b-1fc8-49c3-80a7-ac3469c1504a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm\" (UID: \"c463043b-1fc8-49c3-80a7-ac3469c1504a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.869558 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c463043b-1fc8-49c3-80a7-ac3469c1504a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm\" (UID: \"c463043b-1fc8-49c3-80a7-ac3469c1504a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.870672 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bdf59941-e2f3-4ccf-a700-992e2d3871af-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd\" (UID: \"bdf59941-e2f3-4ccf-a700-992e2d3871af\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.912920 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:01 crc kubenswrapper[4822]: E1124 14:29:01.936662 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(35b857d13ab9796456365780ebdaeda15c41fa49342226557933576f6fec0557): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:01 crc kubenswrapper[4822]: E1124 14:29:01.936759 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(35b857d13ab9796456365780ebdaeda15c41fa49342226557933576f6fec0557): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:01 crc kubenswrapper[4822]: E1124 14:29:01.936785 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(35b857d13ab9796456365780ebdaeda15c41fa49342226557933576f6fec0557): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:01 crc kubenswrapper[4822]: E1124 14:29:01.936844 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators(387c7641-9e5d-4ea0-a86f-c890d68d2903)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators(387c7641-9e5d-4ea0-a86f-c890d68d2903)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(35b857d13ab9796456365780ebdaeda15c41fa49342226557933576f6fec0557): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" podUID="387c7641-9e5d-4ea0-a86f-c890d68d2903" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.946878 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-8h9ks"] Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.947641 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.950254 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-5tmmh" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.958626 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.959001 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz8gk\" (UniqueName: \"kubernetes.io/projected/991aab37-0cca-4f05-a7ff-ff64d2cac3d4-kube-api-access-tz8gk\") pod \"observability-operator-d8bb48f5d-sdtbh\" (UID: \"991aab37-0cca-4f05-a7ff-ff64d2cac3d4\") " pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.959074 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/991aab37-0cca-4f05-a7ff-ff64d2cac3d4-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-sdtbh\" (UID: \"991aab37-0cca-4f05-a7ff-ff64d2cac3d4\") " pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.963062 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/991aab37-0cca-4f05-a7ff-ff64d2cac3d4-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-sdtbh\" (UID: \"991aab37-0cca-4f05-a7ff-ff64d2cac3d4\") " pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:01 crc kubenswrapper[4822]: E1124 14:29:01.978387 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(474a0e0b4131600e8bf1fce574872b4a6f878fed5b7af2207a6ffd1711c8e470): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:01 crc kubenswrapper[4822]: E1124 14:29:01.978475 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(474a0e0b4131600e8bf1fce574872b4a6f878fed5b7af2207a6ffd1711c8e470): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: E1124 14:29:01.978533 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(474a0e0b4131600e8bf1fce574872b4a6f878fed5b7af2207a6ffd1711c8e470): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:01 crc kubenswrapper[4822]: E1124 14:29:01.978600 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators(c463043b-1fc8-49c3-80a7-ac3469c1504a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators(c463043b-1fc8-49c3-80a7-ac3469c1504a)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(474a0e0b4131600e8bf1fce574872b4a6f878fed5b7af2207a6ffd1711c8e470): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" podUID="c463043b-1fc8-49c3-80a7-ac3469c1504a" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.983336 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz8gk\" (UniqueName: \"kubernetes.io/projected/991aab37-0cca-4f05-a7ff-ff64d2cac3d4-kube-api-access-tz8gk\") pod \"observability-operator-d8bb48f5d-sdtbh\" (UID: \"991aab37-0cca-4f05-a7ff-ff64d2cac3d4\") " pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:01 crc kubenswrapper[4822]: I1124 14:29:01.985787 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.013467 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(97f918270fac027fb657ec2a28ef0a27c3b682d809ef555be4abb126bd9bc5e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.013594 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(97f918270fac027fb657ec2a28ef0a27c3b682d809ef555be4abb126bd9bc5e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.013622 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(97f918270fac027fb657ec2a28ef0a27c3b682d809ef555be4abb126bd9bc5e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.013696 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators(bdf59941-e2f3-4ccf-a700-992e2d3871af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators(bdf59941-e2f3-4ccf-a700-992e2d3871af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(97f918270fac027fb657ec2a28ef0a27c3b682d809ef555be4abb126bd9bc5e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" podUID="bdf59941-e2f3-4ccf-a700-992e2d3871af" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.060502 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/394cbee2-1b78-4d5f-b663-317a8d4ff56e-openshift-service-ca\") pod \"perses-operator-5446b9c989-8h9ks\" (UID: \"394cbee2-1b78-4d5f-b663-317a8d4ff56e\") " pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.060592 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxdh6\" (UniqueName: \"kubernetes.io/projected/394cbee2-1b78-4d5f-b663-317a8d4ff56e-kube-api-access-zxdh6\") pod \"perses-operator-5446b9c989-8h9ks\" (UID: \"394cbee2-1b78-4d5f-b663-317a8d4ff56e\") " pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.091917 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.115028 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(94a99dbabc4015a62c8b96de8b3e8080f15a406db4ca4cf71995ba5848d0f8d4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.115122 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(94a99dbabc4015a62c8b96de8b3e8080f15a406db4ca4cf71995ba5848d0f8d4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.115157 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(94a99dbabc4015a62c8b96de8b3e8080f15a406db4ca4cf71995ba5848d0f8d4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.115251 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-sdtbh_openshift-operators(991aab37-0cca-4f05-a7ff-ff64d2cac3d4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-sdtbh_openshift-operators(991aab37-0cca-4f05-a7ff-ff64d2cac3d4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(94a99dbabc4015a62c8b96de8b3e8080f15a406db4ca4cf71995ba5848d0f8d4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" podUID="991aab37-0cca-4f05-a7ff-ff64d2cac3d4" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.162327 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxdh6\" (UniqueName: \"kubernetes.io/projected/394cbee2-1b78-4d5f-b663-317a8d4ff56e-kube-api-access-zxdh6\") pod \"perses-operator-5446b9c989-8h9ks\" (UID: \"394cbee2-1b78-4d5f-b663-317a8d4ff56e\") " pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.162423 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/394cbee2-1b78-4d5f-b663-317a8d4ff56e-openshift-service-ca\") pod \"perses-operator-5446b9c989-8h9ks\" (UID: \"394cbee2-1b78-4d5f-b663-317a8d4ff56e\") " pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.163302 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/394cbee2-1b78-4d5f-b663-317a8d4ff56e-openshift-service-ca\") pod \"perses-operator-5446b9c989-8h9ks\" (UID: \"394cbee2-1b78-4d5f-b663-317a8d4ff56e\") " pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.182602 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxdh6\" (UniqueName: \"kubernetes.io/projected/394cbee2-1b78-4d5f-b663-317a8d4ff56e-kube-api-access-zxdh6\") pod \"perses-operator-5446b9c989-8h9ks\" (UID: \"394cbee2-1b78-4d5f-b663-317a8d4ff56e\") " pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.260840 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.282702 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(7209eaa0a9d57db9214053170d71b4f70e9429a45b8add6c2ff563d864368f13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.282843 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(7209eaa0a9d57db9214053170d71b4f70e9429a45b8add6c2ff563d864368f13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.282929 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(7209eaa0a9d57db9214053170d71b4f70e9429a45b8add6c2ff563d864368f13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.283052 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-8h9ks_openshift-operators(394cbee2-1b78-4d5f-b663-317a8d4ff56e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-8h9ks_openshift-operators(394cbee2-1b78-4d5f-b663-317a8d4ff56e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(7209eaa0a9d57db9214053170d71b4f70e9429a45b8add6c2ff563d864368f13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" podUID="394cbee2-1b78-4d5f-b663-317a8d4ff56e" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.414839 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" event={"ID":"d71763fd-618d-4c6c-b062-52fbffcb68d4","Type":"ContainerStarted","Data":"7f82d93e21d2aed466c482dd596e998248e6b0cc762bed8a5b11ab137543942c"} Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.415459 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.415646 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.415840 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.444252 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.444852 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.452563 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" podStartSLOduration=7.452539451 podStartE2EDuration="7.452539451s" podCreationTimestamp="2025-11-24 14:28:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:29:02.447336765 +0000 UTC m=+579.563977262" watchObservedRunningTime="2025-11-24 14:29:02.452539451 +0000 UTC m=+579.569179928" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.570775 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd"] Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.571223 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.572000 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.578553 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-sdtbh"] Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.578717 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.579224 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.587955 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t"] Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.588139 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.588831 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.595920 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm"] Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.596069 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.596623 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.607111 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(16dd10065440967d2fb9003a5e440b5e520e271d65002a983a43c1e214f6c523): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.607212 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(16dd10065440967d2fb9003a5e440b5e520e271d65002a983a43c1e214f6c523): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.607240 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(16dd10065440967d2fb9003a5e440b5e520e271d65002a983a43c1e214f6c523): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.607307 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators(bdf59941-e2f3-4ccf-a700-992e2d3871af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators(bdf59941-e2f3-4ccf-a700-992e2d3871af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(16dd10065440967d2fb9003a5e440b5e520e271d65002a983a43c1e214f6c523): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" podUID="bdf59941-e2f3-4ccf-a700-992e2d3871af" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.629581 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(e877612dbfb6815d3960f62b18c9783656ebfa76c62c44cb8ab0cd57f34c98d5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.629669 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(e877612dbfb6815d3960f62b18c9783656ebfa76c62c44cb8ab0cd57f34c98d5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.629693 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(e877612dbfb6815d3960f62b18c9783656ebfa76c62c44cb8ab0cd57f34c98d5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.629747 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-sdtbh_openshift-operators(991aab37-0cca-4f05-a7ff-ff64d2cac3d4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-sdtbh_openshift-operators(991aab37-0cca-4f05-a7ff-ff64d2cac3d4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(e877612dbfb6815d3960f62b18c9783656ebfa76c62c44cb8ab0cd57f34c98d5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" podUID="991aab37-0cca-4f05-a7ff-ff64d2cac3d4" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.640820 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-8h9ks"] Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.640962 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: I1124 14:29:02.641649 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.648310 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(96f62f2c5aa856185eaef68c0335dd20c2b3534c3568658c436abdc16c6cdc16): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.648388 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(96f62f2c5aa856185eaef68c0335dd20c2b3534c3568658c436abdc16c6cdc16): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.648417 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(96f62f2c5aa856185eaef68c0335dd20c2b3534c3568658c436abdc16c6cdc16): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.648468 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators(387c7641-9e5d-4ea0-a86f-c890d68d2903)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators(387c7641-9e5d-4ea0-a86f-c890d68d2903)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(96f62f2c5aa856185eaef68c0335dd20c2b3534c3568658c436abdc16c6cdc16): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" podUID="387c7641-9e5d-4ea0-a86f-c890d68d2903" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.665577 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(0cb7f2bc8aed39a7ffe9a6348560e8c579d6bd419457cac32c28f92fb586793a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.665674 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(0cb7f2bc8aed39a7ffe9a6348560e8c579d6bd419457cac32c28f92fb586793a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.665702 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(0cb7f2bc8aed39a7ffe9a6348560e8c579d6bd419457cac32c28f92fb586793a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.665759 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators(c463043b-1fc8-49c3-80a7-ac3469c1504a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators(c463043b-1fc8-49c3-80a7-ac3469c1504a)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(0cb7f2bc8aed39a7ffe9a6348560e8c579d6bd419457cac32c28f92fb586793a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" podUID="c463043b-1fc8-49c3-80a7-ac3469c1504a" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.690842 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(9aa332039ebcc71b32d18cb5c6731908f547b9a1da0a41868d0d511bbe89a30c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.690954 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(9aa332039ebcc71b32d18cb5c6731908f547b9a1da0a41868d0d511bbe89a30c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.690983 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(9aa332039ebcc71b32d18cb5c6731908f547b9a1da0a41868d0d511bbe89a30c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:02 crc kubenswrapper[4822]: E1124 14:29:02.691031 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-8h9ks_openshift-operators(394cbee2-1b78-4d5f-b663-317a8d4ff56e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-8h9ks_openshift-operators(394cbee2-1b78-4d5f-b663-317a8d4ff56e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(9aa332039ebcc71b32d18cb5c6731908f547b9a1da0a41868d0d511bbe89a30c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" podUID="394cbee2-1b78-4d5f-b663-317a8d4ff56e" Nov 24 14:29:08 crc kubenswrapper[4822]: I1124 14:29:08.704810 4822 scope.go:117] "RemoveContainer" containerID="f528354f90b6adf0611b43f5ddea194a67abb6f33c1715df7b8dc36dab6a88f9" Nov 24 14:29:08 crc kubenswrapper[4822]: E1124 14:29:08.705492 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-d6rbt_openshift-multus(92c230aa-aed2-4b07-82bd-98b1d2ffa456)\"" pod="openshift-multus/multus-d6rbt" podUID="92c230aa-aed2-4b07-82bd-98b1d2ffa456" Nov 24 14:29:11 crc kubenswrapper[4822]: I1124 14:29:11.577843 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:29:11 crc kubenswrapper[4822]: I1124 14:29:11.577937 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:29:11 crc kubenswrapper[4822]: I1124 14:29:11.578006 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:29:11 crc kubenswrapper[4822]: I1124 14:29:11.578829 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c48afa8921812e0254372f9525778df4abed8a487a31452c711ba4cfd0103ae3"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:29:11 crc kubenswrapper[4822]: I1124 14:29:11.578928 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://c48afa8921812e0254372f9525778df4abed8a487a31452c711ba4cfd0103ae3" gracePeriod=600 Nov 24 14:29:12 crc kubenswrapper[4822]: I1124 14:29:12.473363 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="c48afa8921812e0254372f9525778df4abed8a487a31452c711ba4cfd0103ae3" exitCode=0 Nov 24 14:29:12 crc kubenswrapper[4822]: I1124 14:29:12.473442 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"c48afa8921812e0254372f9525778df4abed8a487a31452c711ba4cfd0103ae3"} Nov 24 14:29:12 crc kubenswrapper[4822]: I1124 14:29:12.474026 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"065577b1f1af0833feb9bb8a7aed6dd3738bba738b7c830369d5c019d2095f70"} Nov 24 14:29:12 crc kubenswrapper[4822]: I1124 14:29:12.474067 4822 scope.go:117] "RemoveContainer" containerID="7115168ad50bfd88ed09dd9295ac9935a4ada79bd3272f63430ce467dc596e9b" Nov 24 14:29:14 crc kubenswrapper[4822]: I1124 14:29:14.704374 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:14 crc kubenswrapper[4822]: I1124 14:29:14.704385 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:14 crc kubenswrapper[4822]: I1124 14:29:14.705123 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:14 crc kubenswrapper[4822]: I1124 14:29:14.705541 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:14 crc kubenswrapper[4822]: E1124 14:29:14.744728 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(68bb385d4c92ca32651e4acdd78d0d3c070e6d939da526d217cc911a38b59f5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:14 crc kubenswrapper[4822]: E1124 14:29:14.744837 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(68bb385d4c92ca32651e4acdd78d0d3c070e6d939da526d217cc911a38b59f5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:14 crc kubenswrapper[4822]: E1124 14:29:14.744877 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(68bb385d4c92ca32651e4acdd78d0d3c070e6d939da526d217cc911a38b59f5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:14 crc kubenswrapper[4822]: E1124 14:29:14.744956 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-sdtbh_openshift-operators(991aab37-0cca-4f05-a7ff-ff64d2cac3d4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-sdtbh_openshift-operators(991aab37-0cca-4f05-a7ff-ff64d2cac3d4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-sdtbh_openshift-operators_991aab37-0cca-4f05-a7ff-ff64d2cac3d4_0(68bb385d4c92ca32651e4acdd78d0d3c070e6d939da526d217cc911a38b59f5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" podUID="991aab37-0cca-4f05-a7ff-ff64d2cac3d4" Nov 24 14:29:14 crc kubenswrapper[4822]: E1124 14:29:14.749136 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(51adc3d9083e092673646315bb7a79e56ae22e452fe6a43dda170246b1fe3945): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:14 crc kubenswrapper[4822]: E1124 14:29:14.749200 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(51adc3d9083e092673646315bb7a79e56ae22e452fe6a43dda170246b1fe3945): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:14 crc kubenswrapper[4822]: E1124 14:29:14.749253 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(51adc3d9083e092673646315bb7a79e56ae22e452fe6a43dda170246b1fe3945): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:14 crc kubenswrapper[4822]: E1124 14:29:14.749314 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators(387c7641-9e5d-4ea0-a86f-c890d68d2903)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators(387c7641-9e5d-4ea0-a86f-c890d68d2903)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-ql56t_openshift-operators_387c7641-9e5d-4ea0-a86f-c890d68d2903_0(51adc3d9083e092673646315bb7a79e56ae22e452fe6a43dda170246b1fe3945): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" podUID="387c7641-9e5d-4ea0-a86f-c890d68d2903" Nov 24 14:29:16 crc kubenswrapper[4822]: I1124 14:29:16.704381 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:16 crc kubenswrapper[4822]: I1124 14:29:16.704422 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:16 crc kubenswrapper[4822]: I1124 14:29:16.704496 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:16 crc kubenswrapper[4822]: I1124 14:29:16.707030 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:16 crc kubenswrapper[4822]: I1124 14:29:16.707634 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:16 crc kubenswrapper[4822]: I1124 14:29:16.707707 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.764985 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(e62ad745a5fff66e71c37eaf5524cb3f5b4c03980c45ac245c51154fea363b75): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.765054 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(e62ad745a5fff66e71c37eaf5524cb3f5b4c03980c45ac245c51154fea363b75): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.765079 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(e62ad745a5fff66e71c37eaf5524cb3f5b4c03980c45ac245c51154fea363b75): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.765125 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-8h9ks_openshift-operators(394cbee2-1b78-4d5f-b663-317a8d4ff56e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-8h9ks_openshift-operators(394cbee2-1b78-4d5f-b663-317a8d4ff56e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-8h9ks_openshift-operators_394cbee2-1b78-4d5f-b663-317a8d4ff56e_0(e62ad745a5fff66e71c37eaf5524cb3f5b4c03980c45ac245c51154fea363b75): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" podUID="394cbee2-1b78-4d5f-b663-317a8d4ff56e" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.775570 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(ff9cede2fed334af0e29e2d8467428ce37ff414c17316a482018bc8e4c4dffda): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.775942 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(ff9cede2fed334af0e29e2d8467428ce37ff414c17316a482018bc8e4c4dffda): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.775972 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(ff9cede2fed334af0e29e2d8467428ce37ff414c17316a482018bc8e4c4dffda): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.776031 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators(c463043b-1fc8-49c3-80a7-ac3469c1504a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators(c463043b-1fc8-49c3-80a7-ac3469c1504a)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_openshift-operators_c463043b-1fc8-49c3-80a7-ac3469c1504a_0(ff9cede2fed334af0e29e2d8467428ce37ff414c17316a482018bc8e4c4dffda): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" podUID="c463043b-1fc8-49c3-80a7-ac3469c1504a" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.782790 4822 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(935b37e51b9ded87e1c01cd91b8be09c7657520e1f09e673b32fccceed740f3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.782950 4822 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(935b37e51b9ded87e1c01cd91b8be09c7657520e1f09e673b32fccceed740f3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.783043 4822 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(935b37e51b9ded87e1c01cd91b8be09c7657520e1f09e673b32fccceed740f3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:16 crc kubenswrapper[4822]: E1124 14:29:16.783181 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators(bdf59941-e2f3-4ccf-a700-992e2d3871af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators(bdf59941-e2f3-4ccf-a700-992e2d3871af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_openshift-operators_bdf59941-e2f3-4ccf-a700-992e2d3871af_0(935b37e51b9ded87e1c01cd91b8be09c7657520e1f09e673b32fccceed740f3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" podUID="bdf59941-e2f3-4ccf-a700-992e2d3871af" Nov 24 14:29:20 crc kubenswrapper[4822]: I1124 14:29:20.703992 4822 scope.go:117] "RemoveContainer" containerID="f528354f90b6adf0611b43f5ddea194a67abb6f33c1715df7b8dc36dab6a88f9" Nov 24 14:29:21 crc kubenswrapper[4822]: I1124 14:29:21.527594 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-d6rbt_92c230aa-aed2-4b07-82bd-98b1d2ffa456/kube-multus/2.log" Nov 24 14:29:21 crc kubenswrapper[4822]: I1124 14:29:21.527885 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-d6rbt" event={"ID":"92c230aa-aed2-4b07-82bd-98b1d2ffa456","Type":"ContainerStarted","Data":"89f7cdf774cc0a7ec7048cb6a904d54e5251d72d7b3dbbe67c47654be1d0cc9f"} Nov 24 14:29:25 crc kubenswrapper[4822]: I1124 14:29:25.907186 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7m9hl" Nov 24 14:29:26 crc kubenswrapper[4822]: I1124 14:29:26.703894 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:26 crc kubenswrapper[4822]: I1124 14:29:26.704876 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:27 crc kubenswrapper[4822]: I1124 14:29:27.133592 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-sdtbh"] Nov 24 14:29:27 crc kubenswrapper[4822]: W1124 14:29:27.142411 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod991aab37_0cca_4f05_a7ff_ff64d2cac3d4.slice/crio-61e32c4e825a812acb4313998a9ff0b44bdef697b0faddba0a74ef4274c16ebc WatchSource:0}: Error finding container 61e32c4e825a812acb4313998a9ff0b44bdef697b0faddba0a74ef4274c16ebc: Status 404 returned error can't find the container with id 61e32c4e825a812acb4313998a9ff0b44bdef697b0faddba0a74ef4274c16ebc Nov 24 14:29:27 crc kubenswrapper[4822]: I1124 14:29:27.564463 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" event={"ID":"991aab37-0cca-4f05-a7ff-ff64d2cac3d4","Type":"ContainerStarted","Data":"61e32c4e825a812acb4313998a9ff0b44bdef697b0faddba0a74ef4274c16ebc"} Nov 24 14:29:28 crc kubenswrapper[4822]: I1124 14:29:28.703793 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:28 crc kubenswrapper[4822]: I1124 14:29:28.704720 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" Nov 24 14:29:29 crc kubenswrapper[4822]: I1124 14:29:29.171072 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t"] Nov 24 14:29:29 crc kubenswrapper[4822]: W1124 14:29:29.177494 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod387c7641_9e5d_4ea0_a86f_c890d68d2903.slice/crio-b0a6300853cd832ffb6bcff091afae31c0277b4d3bfd2547321d36972842218c WatchSource:0}: Error finding container b0a6300853cd832ffb6bcff091afae31c0277b4d3bfd2547321d36972842218c: Status 404 returned error can't find the container with id b0a6300853cd832ffb6bcff091afae31c0277b4d3bfd2547321d36972842218c Nov 24 14:29:29 crc kubenswrapper[4822]: I1124 14:29:29.578539 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" event={"ID":"387c7641-9e5d-4ea0-a86f-c890d68d2903","Type":"ContainerStarted","Data":"b0a6300853cd832ffb6bcff091afae31c0277b4d3bfd2547321d36972842218c"} Nov 24 14:29:31 crc kubenswrapper[4822]: I1124 14:29:31.707158 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:31 crc kubenswrapper[4822]: I1124 14:29:31.707868 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:31 crc kubenswrapper[4822]: I1124 14:29:31.707177 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:31 crc kubenswrapper[4822]: I1124 14:29:31.708085 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" Nov 24 14:29:31 crc kubenswrapper[4822]: I1124 14:29:31.707414 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:31 crc kubenswrapper[4822]: I1124 14:29:31.708336 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" Nov 24 14:29:36 crc kubenswrapper[4822]: I1124 14:29:36.656808 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd"] Nov 24 14:29:36 crc kubenswrapper[4822]: I1124 14:29:36.938377 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm"] Nov 24 14:29:36 crc kubenswrapper[4822]: I1124 14:29:36.942384 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-8h9ks"] Nov 24 14:29:36 crc kubenswrapper[4822]: W1124 14:29:36.948535 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod394cbee2_1b78_4d5f_b663_317a8d4ff56e.slice/crio-37a86d04d0ad9d203dad0e148be5e9ab1c681f1c5bcc90711e7c35ccbcb06567 WatchSource:0}: Error finding container 37a86d04d0ad9d203dad0e148be5e9ab1c681f1c5bcc90711e7c35ccbcb06567: Status 404 returned error can't find the container with id 37a86d04d0ad9d203dad0e148be5e9ab1c681f1c5bcc90711e7c35ccbcb06567 Nov 24 14:29:36 crc kubenswrapper[4822]: W1124 14:29:36.960099 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc463043b_1fc8_49c3_80a7_ac3469c1504a.slice/crio-decf87838c6f4c3d004cf767df5bf51d75c4ea2574490a8ca18f456259189a03 WatchSource:0}: Error finding container decf87838c6f4c3d004cf767df5bf51d75c4ea2574490a8ca18f456259189a03: Status 404 returned error can't find the container with id decf87838c6f4c3d004cf767df5bf51d75c4ea2574490a8ca18f456259189a03 Nov 24 14:29:37 crc kubenswrapper[4822]: I1124 14:29:37.641472 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" event={"ID":"991aab37-0cca-4f05-a7ff-ff64d2cac3d4","Type":"ContainerStarted","Data":"7c67681b58e01e15159e352d1f4b1be8e51d56b252767eea1e1ac129b726a10a"} Nov 24 14:29:37 crc kubenswrapper[4822]: I1124 14:29:37.641887 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:37 crc kubenswrapper[4822]: I1124 14:29:37.643084 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" event={"ID":"bdf59941-e2f3-4ccf-a700-992e2d3871af","Type":"ContainerStarted","Data":"491b4e54f7b2292dc9ccd426d71f32a1a0b49c5b581f42fc6eea035946879400"} Nov 24 14:29:37 crc kubenswrapper[4822]: I1124 14:29:37.645248 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" Nov 24 14:29:37 crc kubenswrapper[4822]: I1124 14:29:37.645680 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" event={"ID":"394cbee2-1b78-4d5f-b663-317a8d4ff56e","Type":"ContainerStarted","Data":"37a86d04d0ad9d203dad0e148be5e9ab1c681f1c5bcc90711e7c35ccbcb06567"} Nov 24 14:29:37 crc kubenswrapper[4822]: I1124 14:29:37.646985 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" event={"ID":"c463043b-1fc8-49c3-80a7-ac3469c1504a","Type":"ContainerStarted","Data":"decf87838c6f4c3d004cf767df5bf51d75c4ea2574490a8ca18f456259189a03"} Nov 24 14:29:37 crc kubenswrapper[4822]: I1124 14:29:37.648676 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" event={"ID":"387c7641-9e5d-4ea0-a86f-c890d68d2903","Type":"ContainerStarted","Data":"190a0773dc6ce02ce9eab09e82ffc624cb691b15a8c1cbd8f49837218ab3e446"} Nov 24 14:29:37 crc kubenswrapper[4822]: I1124 14:29:37.727940 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-sdtbh" podStartSLOduration=27.398186209 podStartE2EDuration="36.727913512s" podCreationTimestamp="2025-11-24 14:29:01 +0000 UTC" firstStartedPulling="2025-11-24 14:29:27.144858836 +0000 UTC m=+604.261499323" lastFinishedPulling="2025-11-24 14:29:36.474586139 +0000 UTC m=+613.591226626" observedRunningTime="2025-11-24 14:29:37.689099043 +0000 UTC m=+614.805739550" watchObservedRunningTime="2025-11-24 14:29:37.727913512 +0000 UTC m=+614.844554009" Nov 24 14:29:37 crc kubenswrapper[4822]: I1124 14:29:37.767180 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ql56t" podStartSLOduration=29.451357416 podStartE2EDuration="36.767144231s" podCreationTimestamp="2025-11-24 14:29:01 +0000 UTC" firstStartedPulling="2025-11-24 14:29:29.180273539 +0000 UTC m=+606.296914016" lastFinishedPulling="2025-11-24 14:29:36.496060364 +0000 UTC m=+613.612700831" observedRunningTime="2025-11-24 14:29:37.76490159 +0000 UTC m=+614.881542067" watchObservedRunningTime="2025-11-24 14:29:37.767144231 +0000 UTC m=+614.883784718" Nov 24 14:29:39 crc kubenswrapper[4822]: I1124 14:29:39.660408 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" event={"ID":"bdf59941-e2f3-4ccf-a700-992e2d3871af","Type":"ContainerStarted","Data":"35956bce59bacdf8771afb538da77d6f9cf2cb4f9867a7494ee64117d0808b3b"} Nov 24 14:29:39 crc kubenswrapper[4822]: I1124 14:29:39.664155 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" event={"ID":"c463043b-1fc8-49c3-80a7-ac3469c1504a","Type":"ContainerStarted","Data":"02a4b877c3cfa2275bf78cb9bbacd6b1003e7881d6ea842fbdb90843f92f7cda"} Nov 24 14:29:39 crc kubenswrapper[4822]: I1124 14:29:39.689113 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-x96vd" podStartSLOduration=36.236882318 podStartE2EDuration="38.689086417s" podCreationTimestamp="2025-11-24 14:29:01 +0000 UTC" firstStartedPulling="2025-11-24 14:29:36.673308253 +0000 UTC m=+613.789948730" lastFinishedPulling="2025-11-24 14:29:39.125512352 +0000 UTC m=+616.242152829" observedRunningTime="2025-11-24 14:29:39.68757894 +0000 UTC m=+616.804219427" watchObservedRunningTime="2025-11-24 14:29:39.689086417 +0000 UTC m=+616.805726894" Nov 24 14:29:39 crc kubenswrapper[4822]: I1124 14:29:39.712971 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b6d578495-tzppm" podStartSLOduration=36.56732007 podStartE2EDuration="38.712947407s" podCreationTimestamp="2025-11-24 14:29:01 +0000 UTC" firstStartedPulling="2025-11-24 14:29:36.965020712 +0000 UTC m=+614.081661189" lastFinishedPulling="2025-11-24 14:29:39.110648059 +0000 UTC m=+616.227288526" observedRunningTime="2025-11-24 14:29:39.712945817 +0000 UTC m=+616.829586304" watchObservedRunningTime="2025-11-24 14:29:39.712947407 +0000 UTC m=+616.829587894" Nov 24 14:29:41 crc kubenswrapper[4822]: I1124 14:29:41.678679 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" event={"ID":"394cbee2-1b78-4d5f-b663-317a8d4ff56e","Type":"ContainerStarted","Data":"0c208b4225b156d033d55f00f1bd2a017e73ab0192c29105f96c96cefa2706fa"} Nov 24 14:29:41 crc kubenswrapper[4822]: I1124 14:29:41.679248 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:41 crc kubenswrapper[4822]: I1124 14:29:41.699424 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" podStartSLOduration=37.892380831 podStartE2EDuration="40.69940465s" podCreationTimestamp="2025-11-24 14:29:01 +0000 UTC" firstStartedPulling="2025-11-24 14:29:36.951817552 +0000 UTC m=+614.068458029" lastFinishedPulling="2025-11-24 14:29:39.758841371 +0000 UTC m=+616.875481848" observedRunningTime="2025-11-24 14:29:41.695863467 +0000 UTC m=+618.812504024" watchObservedRunningTime="2025-11-24 14:29:41.69940465 +0000 UTC m=+618.816045127" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.224654 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-drv4n"] Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.225854 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-drv4n" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.233518 4822 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wppz9" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.234026 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.243574 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-drv4n"] Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.243619 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.250783 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjnxq\" (UniqueName: \"kubernetes.io/projected/d97bac64-5623-41f7-8338-19ad1704e5de-kube-api-access-rjnxq\") pod \"cert-manager-cainjector-7f985d654d-drv4n\" (UID: \"d97bac64-5623-41f7-8338-19ad1704e5de\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-drv4n" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.251050 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-jnzqf"] Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.251820 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-jnzqf" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.262177 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hxldz"] Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.262894 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.264958 4822 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mz5kq" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.269084 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-jnzqf"] Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.277321 4822 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-h9wxw" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.277503 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hxldz"] Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.352006 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjnxq\" (UniqueName: \"kubernetes.io/projected/d97bac64-5623-41f7-8338-19ad1704e5de-kube-api-access-rjnxq\") pod \"cert-manager-cainjector-7f985d654d-drv4n\" (UID: \"d97bac64-5623-41f7-8338-19ad1704e5de\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-drv4n" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.352106 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgvpp\" (UniqueName: \"kubernetes.io/projected/640c8f8a-b96f-49ba-afd5-c4c80ac909c0-kube-api-access-cgvpp\") pod \"cert-manager-webhook-5655c58dd6-hxldz\" (UID: \"640c8f8a-b96f-49ba-afd5-c4c80ac909c0\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.352176 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w952z\" (UniqueName: \"kubernetes.io/projected/e73af35b-1628-45a3-8a96-1e20b8bca297-kube-api-access-w952z\") pod \"cert-manager-5b446d88c5-jnzqf\" (UID: \"e73af35b-1628-45a3-8a96-1e20b8bca297\") " pod="cert-manager/cert-manager-5b446d88c5-jnzqf" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.372076 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjnxq\" (UniqueName: \"kubernetes.io/projected/d97bac64-5623-41f7-8338-19ad1704e5de-kube-api-access-rjnxq\") pod \"cert-manager-cainjector-7f985d654d-drv4n\" (UID: \"d97bac64-5623-41f7-8338-19ad1704e5de\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-drv4n" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.453160 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgvpp\" (UniqueName: \"kubernetes.io/projected/640c8f8a-b96f-49ba-afd5-c4c80ac909c0-kube-api-access-cgvpp\") pod \"cert-manager-webhook-5655c58dd6-hxldz\" (UID: \"640c8f8a-b96f-49ba-afd5-c4c80ac909c0\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.453234 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w952z\" (UniqueName: \"kubernetes.io/projected/e73af35b-1628-45a3-8a96-1e20b8bca297-kube-api-access-w952z\") pod \"cert-manager-5b446d88c5-jnzqf\" (UID: \"e73af35b-1628-45a3-8a96-1e20b8bca297\") " pod="cert-manager/cert-manager-5b446d88c5-jnzqf" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.470416 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgvpp\" (UniqueName: \"kubernetes.io/projected/640c8f8a-b96f-49ba-afd5-c4c80ac909c0-kube-api-access-cgvpp\") pod \"cert-manager-webhook-5655c58dd6-hxldz\" (UID: \"640c8f8a-b96f-49ba-afd5-c4c80ac909c0\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.475038 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w952z\" (UniqueName: \"kubernetes.io/projected/e73af35b-1628-45a3-8a96-1e20b8bca297-kube-api-access-w952z\") pod \"cert-manager-5b446d88c5-jnzqf\" (UID: \"e73af35b-1628-45a3-8a96-1e20b8bca297\") " pod="cert-manager/cert-manager-5b446d88c5-jnzqf" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.540130 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-drv4n" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.564139 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-jnzqf" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.575940 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" Nov 24 14:29:47 crc kubenswrapper[4822]: I1124 14:29:47.840562 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-drv4n"] Nov 24 14:29:48 crc kubenswrapper[4822]: I1124 14:29:48.105007 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-jnzqf"] Nov 24 14:29:48 crc kubenswrapper[4822]: I1124 14:29:48.112963 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hxldz"] Nov 24 14:29:48 crc kubenswrapper[4822]: W1124 14:29:48.121864 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod640c8f8a_b96f_49ba_afd5_c4c80ac909c0.slice/crio-7daf2e88f2d4d7b895ea58382f034d0d5d2e2fef06013c7c566f00961bd374e8 WatchSource:0}: Error finding container 7daf2e88f2d4d7b895ea58382f034d0d5d2e2fef06013c7c566f00961bd374e8: Status 404 returned error can't find the container with id 7daf2e88f2d4d7b895ea58382f034d0d5d2e2fef06013c7c566f00961bd374e8 Nov 24 14:29:48 crc kubenswrapper[4822]: I1124 14:29:48.751337 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" event={"ID":"640c8f8a-b96f-49ba-afd5-c4c80ac909c0","Type":"ContainerStarted","Data":"7daf2e88f2d4d7b895ea58382f034d0d5d2e2fef06013c7c566f00961bd374e8"} Nov 24 14:29:48 crc kubenswrapper[4822]: I1124 14:29:48.753291 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-drv4n" event={"ID":"d97bac64-5623-41f7-8338-19ad1704e5de","Type":"ContainerStarted","Data":"a4e53003c832a26e9a32ff6e66cbb7674458a21b52584b4bfc6c9405c945b1ec"} Nov 24 14:29:48 crc kubenswrapper[4822]: I1124 14:29:48.755534 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-jnzqf" event={"ID":"e73af35b-1628-45a3-8a96-1e20b8bca297","Type":"ContainerStarted","Data":"7a3557af352f8178834238063eca0ec5619369ce4ef13db3bc22219055c7121f"} Nov 24 14:29:51 crc kubenswrapper[4822]: I1124 14:29:51.777262 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" event={"ID":"640c8f8a-b96f-49ba-afd5-c4c80ac909c0","Type":"ContainerStarted","Data":"960102324b82d75533d224856d9095f575e74721176735d7de6015e3af5a89b1"} Nov 24 14:29:51 crc kubenswrapper[4822]: I1124 14:29:51.778266 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" Nov 24 14:29:51 crc kubenswrapper[4822]: I1124 14:29:51.779270 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-drv4n" event={"ID":"d97bac64-5623-41f7-8338-19ad1704e5de","Type":"ContainerStarted","Data":"c7f60010ea0a8221236f9bafcd8021ce58e9f975c4d3b6e6539b6fb30d9d7e9c"} Nov 24 14:29:51 crc kubenswrapper[4822]: I1124 14:29:51.783015 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-jnzqf" event={"ID":"e73af35b-1628-45a3-8a96-1e20b8bca297","Type":"ContainerStarted","Data":"3ca71d31151665f17ed403ff407eaa83ecd157806490f68392df5c7de796075b"} Nov 24 14:29:51 crc kubenswrapper[4822]: I1124 14:29:51.795088 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" podStartSLOduration=1.396535014 podStartE2EDuration="4.795064798s" podCreationTimestamp="2025-11-24 14:29:47 +0000 UTC" firstStartedPulling="2025-11-24 14:29:48.125300408 +0000 UTC m=+625.241940925" lastFinishedPulling="2025-11-24 14:29:51.523830222 +0000 UTC m=+628.640470709" observedRunningTime="2025-11-24 14:29:51.793799268 +0000 UTC m=+628.910439765" watchObservedRunningTime="2025-11-24 14:29:51.795064798 +0000 UTC m=+628.911705275" Nov 24 14:29:51 crc kubenswrapper[4822]: I1124 14:29:51.810220 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-drv4n" podStartSLOduration=1.1374874560000001 podStartE2EDuration="4.8101881s" podCreationTimestamp="2025-11-24 14:29:47 +0000 UTC" firstStartedPulling="2025-11-24 14:29:47.850692914 +0000 UTC m=+624.967333401" lastFinishedPulling="2025-11-24 14:29:51.523393548 +0000 UTC m=+628.640034045" observedRunningTime="2025-11-24 14:29:51.807922628 +0000 UTC m=+628.924563105" watchObservedRunningTime="2025-11-24 14:29:51.8101881 +0000 UTC m=+628.926828577" Nov 24 14:29:51 crc kubenswrapper[4822]: I1124 14:29:51.866389 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-jnzqf" podStartSLOduration=1.468646963 podStartE2EDuration="4.866365961s" podCreationTimestamp="2025-11-24 14:29:47 +0000 UTC" firstStartedPulling="2025-11-24 14:29:48.113319566 +0000 UTC m=+625.229960083" lastFinishedPulling="2025-11-24 14:29:51.511038584 +0000 UTC m=+628.627679081" observedRunningTime="2025-11-24 14:29:51.862304422 +0000 UTC m=+628.978944909" watchObservedRunningTime="2025-11-24 14:29:51.866365961 +0000 UTC m=+628.983006448" Nov 24 14:29:52 crc kubenswrapper[4822]: I1124 14:29:52.264963 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-8h9ks" Nov 24 14:29:57 crc kubenswrapper[4822]: I1124 14:29:57.578435 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-hxldz" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.144637 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv"] Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.146577 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.149627 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.150910 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.161103 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv"] Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.228720 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd909522-d5b8-4024-a3c5-f1d04821a578-secret-volume\") pod \"collect-profiles-29399910-ljbfv\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.228801 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd909522-d5b8-4024-a3c5-f1d04821a578-config-volume\") pod \"collect-profiles-29399910-ljbfv\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.228863 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsdxp\" (UniqueName: \"kubernetes.io/projected/fd909522-d5b8-4024-a3c5-f1d04821a578-kube-api-access-bsdxp\") pod \"collect-profiles-29399910-ljbfv\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.330461 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd909522-d5b8-4024-a3c5-f1d04821a578-secret-volume\") pod \"collect-profiles-29399910-ljbfv\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.330533 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd909522-d5b8-4024-a3c5-f1d04821a578-config-volume\") pod \"collect-profiles-29399910-ljbfv\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.330588 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsdxp\" (UniqueName: \"kubernetes.io/projected/fd909522-d5b8-4024-a3c5-f1d04821a578-kube-api-access-bsdxp\") pod \"collect-profiles-29399910-ljbfv\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.332137 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd909522-d5b8-4024-a3c5-f1d04821a578-config-volume\") pod \"collect-profiles-29399910-ljbfv\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.336356 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd909522-d5b8-4024-a3c5-f1d04821a578-secret-volume\") pod \"collect-profiles-29399910-ljbfv\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.345172 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsdxp\" (UniqueName: \"kubernetes.io/projected/fd909522-d5b8-4024-a3c5-f1d04821a578-kube-api-access-bsdxp\") pod \"collect-profiles-29399910-ljbfv\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.476143 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.714616 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv"] Nov 24 14:30:00 crc kubenswrapper[4822]: I1124 14:30:00.837050 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" event={"ID":"fd909522-d5b8-4024-a3c5-f1d04821a578","Type":"ContainerStarted","Data":"4b417d4a44fcfaea7963bfd9c6dbda70bd7b595372d1bd678d47ed320731548b"} Nov 24 14:30:01 crc kubenswrapper[4822]: I1124 14:30:01.844677 4822 generic.go:334] "Generic (PLEG): container finished" podID="fd909522-d5b8-4024-a3c5-f1d04821a578" containerID="c9de05de335c3cd8837277fbae4a5826a2f3c0721dd9e5338f7dfaaaa3e5ea71" exitCode=0 Nov 24 14:30:01 crc kubenswrapper[4822]: I1124 14:30:01.844990 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" event={"ID":"fd909522-d5b8-4024-a3c5-f1d04821a578","Type":"ContainerDied","Data":"c9de05de335c3cd8837277fbae4a5826a2f3c0721dd9e5338f7dfaaaa3e5ea71"} Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.123097 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.167130 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsdxp\" (UniqueName: \"kubernetes.io/projected/fd909522-d5b8-4024-a3c5-f1d04821a578-kube-api-access-bsdxp\") pod \"fd909522-d5b8-4024-a3c5-f1d04821a578\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.167183 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd909522-d5b8-4024-a3c5-f1d04821a578-secret-volume\") pod \"fd909522-d5b8-4024-a3c5-f1d04821a578\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.167327 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd909522-d5b8-4024-a3c5-f1d04821a578-config-volume\") pod \"fd909522-d5b8-4024-a3c5-f1d04821a578\" (UID: \"fd909522-d5b8-4024-a3c5-f1d04821a578\") " Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.168657 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd909522-d5b8-4024-a3c5-f1d04821a578-config-volume" (OuterVolumeSpecName: "config-volume") pod "fd909522-d5b8-4024-a3c5-f1d04821a578" (UID: "fd909522-d5b8-4024-a3c5-f1d04821a578"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.188475 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd909522-d5b8-4024-a3c5-f1d04821a578-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fd909522-d5b8-4024-a3c5-f1d04821a578" (UID: "fd909522-d5b8-4024-a3c5-f1d04821a578"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.188581 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd909522-d5b8-4024-a3c5-f1d04821a578-kube-api-access-bsdxp" (OuterVolumeSpecName: "kube-api-access-bsdxp") pod "fd909522-d5b8-4024-a3c5-f1d04821a578" (UID: "fd909522-d5b8-4024-a3c5-f1d04821a578"). InnerVolumeSpecName "kube-api-access-bsdxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.269413 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsdxp\" (UniqueName: \"kubernetes.io/projected/fd909522-d5b8-4024-a3c5-f1d04821a578-kube-api-access-bsdxp\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.269468 4822 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd909522-d5b8-4024-a3c5-f1d04821a578-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.269481 4822 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd909522-d5b8-4024-a3c5-f1d04821a578-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.860502 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" event={"ID":"fd909522-d5b8-4024-a3c5-f1d04821a578","Type":"ContainerDied","Data":"4b417d4a44fcfaea7963bfd9c6dbda70bd7b595372d1bd678d47ed320731548b"} Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.860578 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b417d4a44fcfaea7963bfd9c6dbda70bd7b595372d1bd678d47ed320731548b" Nov 24 14:30:03 crc kubenswrapper[4822]: I1124 14:30:03.860618 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv" Nov 24 14:30:28 crc kubenswrapper[4822]: I1124 14:30:28.926645 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx"] Nov 24 14:30:28 crc kubenswrapper[4822]: E1124 14:30:28.927232 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd909522-d5b8-4024-a3c5-f1d04821a578" containerName="collect-profiles" Nov 24 14:30:28 crc kubenswrapper[4822]: I1124 14:30:28.927244 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd909522-d5b8-4024-a3c5-f1d04821a578" containerName="collect-profiles" Nov 24 14:30:28 crc kubenswrapper[4822]: I1124 14:30:28.927353 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd909522-d5b8-4024-a3c5-f1d04821a578" containerName="collect-profiles" Nov 24 14:30:28 crc kubenswrapper[4822]: I1124 14:30:28.928085 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:28 crc kubenswrapper[4822]: I1124 14:30:28.936607 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx"] Nov 24 14:30:28 crc kubenswrapper[4822]: I1124 14:30:28.937372 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.015376 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-util\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.015476 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-bundle\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.015513 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qhtv\" (UniqueName: \"kubernetes.io/projected/7f24b1dd-5e33-4c52-af2d-43a4c9849225-kube-api-access-9qhtv\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.116752 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-bundle\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.116833 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qhtv\" (UniqueName: \"kubernetes.io/projected/7f24b1dd-5e33-4c52-af2d-43a4c9849225-kube-api-access-9qhtv\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.116931 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-util\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.117352 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-bundle\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.117594 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-util\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.138101 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qhtv\" (UniqueName: \"kubernetes.io/projected/7f24b1dd-5e33-4c52-af2d-43a4c9849225-kube-api-access-9qhtv\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.243558 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:29 crc kubenswrapper[4822]: I1124 14:30:29.466410 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx"] Nov 24 14:30:30 crc kubenswrapper[4822]: I1124 14:30:30.059996 4822 generic.go:334] "Generic (PLEG): container finished" podID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerID="cf36297b10c24fa24d2d03c6f0043e8a5974f184ee2a7f07f6714bb866e80907" exitCode=0 Nov 24 14:30:30 crc kubenswrapper[4822]: I1124 14:30:30.060090 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" event={"ID":"7f24b1dd-5e33-4c52-af2d-43a4c9849225","Type":"ContainerDied","Data":"cf36297b10c24fa24d2d03c6f0043e8a5974f184ee2a7f07f6714bb866e80907"} Nov 24 14:30:30 crc kubenswrapper[4822]: I1124 14:30:30.060550 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" event={"ID":"7f24b1dd-5e33-4c52-af2d-43a4c9849225","Type":"ContainerStarted","Data":"f3c66f15a1f5f91b1ba7d20a19baa70e69e90feb20e26ebfd423de8748e1750a"} Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.673535 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.683555 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.693414 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.693437 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.696831 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.750147 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-566816fd-385b-49e6-8de6-4744b0b480ef\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-566816fd-385b-49e6-8de6-4744b0b480ef\") pod \"minio\" (UID: \"1e94d615-d766-49e0-9c59-6834625a37f7\") " pod="minio-dev/minio" Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.750245 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kphw\" (UniqueName: \"kubernetes.io/projected/1e94d615-d766-49e0-9c59-6834625a37f7-kube-api-access-9kphw\") pod \"minio\" (UID: \"1e94d615-d766-49e0-9c59-6834625a37f7\") " pod="minio-dev/minio" Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.851097 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-566816fd-385b-49e6-8de6-4744b0b480ef\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-566816fd-385b-49e6-8de6-4744b0b480ef\") pod \"minio\" (UID: \"1e94d615-d766-49e0-9c59-6834625a37f7\") " pod="minio-dev/minio" Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.851185 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kphw\" (UniqueName: \"kubernetes.io/projected/1e94d615-d766-49e0-9c59-6834625a37f7-kube-api-access-9kphw\") pod \"minio\" (UID: \"1e94d615-d766-49e0-9c59-6834625a37f7\") " pod="minio-dev/minio" Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.855508 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.855554 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-566816fd-385b-49e6-8de6-4744b0b480ef\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-566816fd-385b-49e6-8de6-4744b0b480ef\") pod \"minio\" (UID: \"1e94d615-d766-49e0-9c59-6834625a37f7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d301008a054bdd5a4b1409c98bf6a99e91483dfb22b9012a7be4b6f7b10e24dd/globalmount\"" pod="minio-dev/minio" Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.876070 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kphw\" (UniqueName: \"kubernetes.io/projected/1e94d615-d766-49e0-9c59-6834625a37f7-kube-api-access-9kphw\") pod \"minio\" (UID: \"1e94d615-d766-49e0-9c59-6834625a37f7\") " pod="minio-dev/minio" Nov 24 14:30:31 crc kubenswrapper[4822]: I1124 14:30:31.881882 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-566816fd-385b-49e6-8de6-4744b0b480ef\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-566816fd-385b-49e6-8de6-4744b0b480ef\") pod \"minio\" (UID: \"1e94d615-d766-49e0-9c59-6834625a37f7\") " pod="minio-dev/minio" Nov 24 14:30:32 crc kubenswrapper[4822]: I1124 14:30:32.026431 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 24 14:30:32 crc kubenswrapper[4822]: I1124 14:30:32.348829 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 24 14:30:32 crc kubenswrapper[4822]: W1124 14:30:32.353238 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e94d615_d766_49e0_9c59_6834625a37f7.slice/crio-90fc2856dc9f34dc897dee25dee4ffb26cf2142bdc11354ddb7a4c55beba50e1 WatchSource:0}: Error finding container 90fc2856dc9f34dc897dee25dee4ffb26cf2142bdc11354ddb7a4c55beba50e1: Status 404 returned error can't find the container with id 90fc2856dc9f34dc897dee25dee4ffb26cf2142bdc11354ddb7a4c55beba50e1 Nov 24 14:30:33 crc kubenswrapper[4822]: I1124 14:30:33.083886 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"1e94d615-d766-49e0-9c59-6834625a37f7","Type":"ContainerStarted","Data":"90fc2856dc9f34dc897dee25dee4ffb26cf2142bdc11354ddb7a4c55beba50e1"} Nov 24 14:30:33 crc kubenswrapper[4822]: I1124 14:30:33.087640 4822 generic.go:334] "Generic (PLEG): container finished" podID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerID="ce202a457475adf28c17b9de5b9f4cc71d480585c5a139349492968ba9b88456" exitCode=0 Nov 24 14:30:33 crc kubenswrapper[4822]: I1124 14:30:33.087670 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" event={"ID":"7f24b1dd-5e33-4c52-af2d-43a4c9849225","Type":"ContainerDied","Data":"ce202a457475adf28c17b9de5b9f4cc71d480585c5a139349492968ba9b88456"} Nov 24 14:30:34 crc kubenswrapper[4822]: I1124 14:30:34.094351 4822 generic.go:334] "Generic (PLEG): container finished" podID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerID="756c54677e76d4a36fe2834bef4be8332c16531e48c01f329b98af0b4f560f39" exitCode=0 Nov 24 14:30:34 crc kubenswrapper[4822]: I1124 14:30:34.094994 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" event={"ID":"7f24b1dd-5e33-4c52-af2d-43a4c9849225","Type":"ContainerDied","Data":"756c54677e76d4a36fe2834bef4be8332c16531e48c01f329b98af0b4f560f39"} Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.589716 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.610699 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-bundle\") pod \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.610781 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-util\") pod \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.610834 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qhtv\" (UniqueName: \"kubernetes.io/projected/7f24b1dd-5e33-4c52-af2d-43a4c9849225-kube-api-access-9qhtv\") pod \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\" (UID: \"7f24b1dd-5e33-4c52-af2d-43a4c9849225\") " Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.611973 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-bundle" (OuterVolumeSpecName: "bundle") pod "7f24b1dd-5e33-4c52-af2d-43a4c9849225" (UID: "7f24b1dd-5e33-4c52-af2d-43a4c9849225"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.618970 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f24b1dd-5e33-4c52-af2d-43a4c9849225-kube-api-access-9qhtv" (OuterVolumeSpecName: "kube-api-access-9qhtv") pod "7f24b1dd-5e33-4c52-af2d-43a4c9849225" (UID: "7f24b1dd-5e33-4c52-af2d-43a4c9849225"). InnerVolumeSpecName "kube-api-access-9qhtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.628085 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-util" (OuterVolumeSpecName: "util") pod "7f24b1dd-5e33-4c52-af2d-43a4c9849225" (UID: "7f24b1dd-5e33-4c52-af2d-43a4c9849225"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.712705 4822 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.712738 4822 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f24b1dd-5e33-4c52-af2d-43a4c9849225-util\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:35 crc kubenswrapper[4822]: I1124 14:30:35.712751 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qhtv\" (UniqueName: \"kubernetes.io/projected/7f24b1dd-5e33-4c52-af2d-43a4c9849225-kube-api-access-9qhtv\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:36 crc kubenswrapper[4822]: I1124 14:30:36.113104 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"1e94d615-d766-49e0-9c59-6834625a37f7","Type":"ContainerStarted","Data":"4c6fc490f450c65f21834388e03c264d8c713aef07f092dae83488062aaafda3"} Nov 24 14:30:36 crc kubenswrapper[4822]: I1124 14:30:36.117087 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" event={"ID":"7f24b1dd-5e33-4c52-af2d-43a4c9849225","Type":"ContainerDied","Data":"f3c66f15a1f5f91b1ba7d20a19baa70e69e90feb20e26ebfd423de8748e1750a"} Nov 24 14:30:36 crc kubenswrapper[4822]: I1124 14:30:36.117353 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3c66f15a1f5f91b1ba7d20a19baa70e69e90feb20e26ebfd423de8748e1750a" Nov 24 14:30:36 crc kubenswrapper[4822]: I1124 14:30:36.117173 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx" Nov 24 14:30:36 crc kubenswrapper[4822]: I1124 14:30:36.132096 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=3.671125165 podStartE2EDuration="7.132072388s" podCreationTimestamp="2025-11-24 14:30:29 +0000 UTC" firstStartedPulling="2025-11-24 14:30:32.356125337 +0000 UTC m=+669.472765824" lastFinishedPulling="2025-11-24 14:30:35.81707257 +0000 UTC m=+672.933713047" observedRunningTime="2025-11-24 14:30:36.130686484 +0000 UTC m=+673.247326961" watchObservedRunningTime="2025-11-24 14:30:36.132072388 +0000 UTC m=+673.248712875" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.749273 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5"] Nov 24 14:30:41 crc kubenswrapper[4822]: E1124 14:30:41.749747 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerName="extract" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.749761 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerName="extract" Nov 24 14:30:41 crc kubenswrapper[4822]: E1124 14:30:41.749771 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerName="pull" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.749777 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerName="pull" Nov 24 14:30:41 crc kubenswrapper[4822]: E1124 14:30:41.749796 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerName="util" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.749802 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerName="util" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.749897 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f24b1dd-5e33-4c52-af2d-43a4c9849225" containerName="extract" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.750681 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.754365 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.793617 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.793717 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fftrh\" (UniqueName: \"kubernetes.io/projected/20e971ff-ea2d-4652-901b-f744c846972b-kube-api-access-fftrh\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.793741 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.824918 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5"] Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.894658 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fftrh\" (UniqueName: \"kubernetes.io/projected/20e971ff-ea2d-4652-901b-f744c846972b-kube-api-access-fftrh\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.895000 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.895068 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.895541 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.895850 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:41 crc kubenswrapper[4822]: I1124 14:30:41.937426 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fftrh\" (UniqueName: \"kubernetes.io/projected/20e971ff-ea2d-4652-901b-f744c846972b-kube-api-access-fftrh\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:42 crc kubenswrapper[4822]: I1124 14:30:42.072295 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:42 crc kubenswrapper[4822]: I1124 14:30:42.347944 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5"] Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.163101 4822 generic.go:334] "Generic (PLEG): container finished" podID="20e971ff-ea2d-4652-901b-f744c846972b" containerID="2c8e5446573a1dabe3c6602ef8a83b44ea9c1f7c94dcca4f8837120f783816d5" exitCode=0 Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.163179 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" event={"ID":"20e971ff-ea2d-4652-901b-f744c846972b","Type":"ContainerDied","Data":"2c8e5446573a1dabe3c6602ef8a83b44ea9c1f7c94dcca4f8837120f783816d5"} Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.163565 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" event={"ID":"20e971ff-ea2d-4652-901b-f744c846972b","Type":"ContainerStarted","Data":"17afd22230c96225c9cd5f697d68eed83b6df21637b48d5a85d5c016f2719a85"} Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.595636 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv"] Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.596530 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.601712 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.601910 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.602031 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.611900 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.612037 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.615535 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-cqvlq" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.619696 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/cef545df-04fd-49bd-b98a-bcb23fd874f0-manager-config\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.619747 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cef545df-04fd-49bd-b98a-bcb23fd874f0-apiservice-cert\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.619779 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfjvw\" (UniqueName: \"kubernetes.io/projected/cef545df-04fd-49bd-b98a-bcb23fd874f0-kube-api-access-jfjvw\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.619797 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cef545df-04fd-49bd-b98a-bcb23fd874f0-webhook-cert\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.619822 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cef545df-04fd-49bd-b98a-bcb23fd874f0-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.643008 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv"] Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.722538 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cef545df-04fd-49bd-b98a-bcb23fd874f0-apiservice-cert\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.722594 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfjvw\" (UniqueName: \"kubernetes.io/projected/cef545df-04fd-49bd-b98a-bcb23fd874f0-kube-api-access-jfjvw\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.722612 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cef545df-04fd-49bd-b98a-bcb23fd874f0-webhook-cert\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.722641 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cef545df-04fd-49bd-b98a-bcb23fd874f0-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.722690 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/cef545df-04fd-49bd-b98a-bcb23fd874f0-manager-config\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.723601 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/cef545df-04fd-49bd-b98a-bcb23fd874f0-manager-config\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.728860 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cef545df-04fd-49bd-b98a-bcb23fd874f0-webhook-cert\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.739002 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cef545df-04fd-49bd-b98a-bcb23fd874f0-apiservice-cert\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.739836 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cef545df-04fd-49bd-b98a-bcb23fd874f0-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.747256 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfjvw\" (UniqueName: \"kubernetes.io/projected/cef545df-04fd-49bd-b98a-bcb23fd874f0-kube-api-access-jfjvw\") pod \"loki-operator-controller-manager-db9f8c7d4-kz5dv\" (UID: \"cef545df-04fd-49bd-b98a-bcb23fd874f0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:43 crc kubenswrapper[4822]: I1124 14:30:43.913958 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:44 crc kubenswrapper[4822]: I1124 14:30:44.427723 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv"] Nov 24 14:30:44 crc kubenswrapper[4822]: W1124 14:30:44.464222 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcef545df_04fd_49bd_b98a_bcb23fd874f0.slice/crio-b9ef934a5723b19348f2d780a7ab229b97b635c2aa46da5982b8828edd054dc4 WatchSource:0}: Error finding container b9ef934a5723b19348f2d780a7ab229b97b635c2aa46da5982b8828edd054dc4: Status 404 returned error can't find the container with id b9ef934a5723b19348f2d780a7ab229b97b635c2aa46da5982b8828edd054dc4 Nov 24 14:30:45 crc kubenswrapper[4822]: I1124 14:30:45.183931 4822 generic.go:334] "Generic (PLEG): container finished" podID="20e971ff-ea2d-4652-901b-f744c846972b" containerID="b6095254f542cb111678dacae649356042cee1130fd4bc29b3986d999871361c" exitCode=0 Nov 24 14:30:45 crc kubenswrapper[4822]: I1124 14:30:45.184011 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" event={"ID":"20e971ff-ea2d-4652-901b-f744c846972b","Type":"ContainerDied","Data":"b6095254f542cb111678dacae649356042cee1130fd4bc29b3986d999871361c"} Nov 24 14:30:45 crc kubenswrapper[4822]: I1124 14:30:45.186289 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" event={"ID":"cef545df-04fd-49bd-b98a-bcb23fd874f0","Type":"ContainerStarted","Data":"b9ef934a5723b19348f2d780a7ab229b97b635c2aa46da5982b8828edd054dc4"} Nov 24 14:30:46 crc kubenswrapper[4822]: I1124 14:30:46.200641 4822 generic.go:334] "Generic (PLEG): container finished" podID="20e971ff-ea2d-4652-901b-f744c846972b" containerID="0d98b8912e6664f38d112dbde0997d7b5c2039c380a2409303cf9f4438e3d757" exitCode=0 Nov 24 14:30:46 crc kubenswrapper[4822]: I1124 14:30:46.200726 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" event={"ID":"20e971ff-ea2d-4652-901b-f744c846972b","Type":"ContainerDied","Data":"0d98b8912e6664f38d112dbde0997d7b5c2039c380a2409303cf9f4438e3d757"} Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.478995 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.588657 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-util\") pod \"20e971ff-ea2d-4652-901b-f744c846972b\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.588773 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fftrh\" (UniqueName: \"kubernetes.io/projected/20e971ff-ea2d-4652-901b-f744c846972b-kube-api-access-fftrh\") pod \"20e971ff-ea2d-4652-901b-f744c846972b\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.588905 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-bundle\") pod \"20e971ff-ea2d-4652-901b-f744c846972b\" (UID: \"20e971ff-ea2d-4652-901b-f744c846972b\") " Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.590736 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-bundle" (OuterVolumeSpecName: "bundle") pod "20e971ff-ea2d-4652-901b-f744c846972b" (UID: "20e971ff-ea2d-4652-901b-f744c846972b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.610816 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20e971ff-ea2d-4652-901b-f744c846972b-kube-api-access-fftrh" (OuterVolumeSpecName: "kube-api-access-fftrh") pod "20e971ff-ea2d-4652-901b-f744c846972b" (UID: "20e971ff-ea2d-4652-901b-f744c846972b"). InnerVolumeSpecName "kube-api-access-fftrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.613365 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-util" (OuterVolumeSpecName: "util") pod "20e971ff-ea2d-4652-901b-f744c846972b" (UID: "20e971ff-ea2d-4652-901b-f744c846972b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.690579 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fftrh\" (UniqueName: \"kubernetes.io/projected/20e971ff-ea2d-4652-901b-f744c846972b-kube-api-access-fftrh\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.690636 4822 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:48 crc kubenswrapper[4822]: I1124 14:30:48.690652 4822 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20e971ff-ea2d-4652-901b-f744c846972b-util\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:49 crc kubenswrapper[4822]: I1124 14:30:49.222459 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" event={"ID":"20e971ff-ea2d-4652-901b-f744c846972b","Type":"ContainerDied","Data":"17afd22230c96225c9cd5f697d68eed83b6df21637b48d5a85d5c016f2719a85"} Nov 24 14:30:49 crc kubenswrapper[4822]: I1124 14:30:49.222828 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17afd22230c96225c9cd5f697d68eed83b6df21637b48d5a85d5c016f2719a85" Nov 24 14:30:49 crc kubenswrapper[4822]: I1124 14:30:49.222878 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5" Nov 24 14:30:50 crc kubenswrapper[4822]: I1124 14:30:50.229486 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" event={"ID":"cef545df-04fd-49bd-b98a-bcb23fd874f0","Type":"ContainerStarted","Data":"b3019f78e7e841c011f036453cc66d3dc3baa6e528d7bf363c74918f3439652e"} Nov 24 14:30:56 crc kubenswrapper[4822]: I1124 14:30:56.274259 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" event={"ID":"cef545df-04fd-49bd-b98a-bcb23fd874f0","Type":"ContainerStarted","Data":"e83975640eaae83d3befe7ed52d2bd3db80a133cbc853da7f711d913f8e05eba"} Nov 24 14:30:56 crc kubenswrapper[4822]: I1124 14:30:56.277415 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:56 crc kubenswrapper[4822]: I1124 14:30:56.283321 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" Nov 24 14:30:56 crc kubenswrapper[4822]: I1124 14:30:56.312029 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-db9f8c7d4-kz5dv" podStartSLOduration=2.08442064 podStartE2EDuration="13.311997863s" podCreationTimestamp="2025-11-24 14:30:43 +0000 UTC" firstStartedPulling="2025-11-24 14:30:44.467372657 +0000 UTC m=+681.584013144" lastFinishedPulling="2025-11-24 14:30:55.69494989 +0000 UTC m=+692.811590367" observedRunningTime="2025-11-24 14:30:56.307173409 +0000 UTC m=+693.423813896" watchObservedRunningTime="2025-11-24 14:30:56.311997863 +0000 UTC m=+693.428638360" Nov 24 14:31:11 crc kubenswrapper[4822]: I1124 14:31:11.578355 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:31:11 crc kubenswrapper[4822]: I1124 14:31:11.578977 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.295322 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8"] Nov 24 14:31:20 crc kubenswrapper[4822]: E1124 14:31:20.296271 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e971ff-ea2d-4652-901b-f744c846972b" containerName="extract" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.296291 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e971ff-ea2d-4652-901b-f744c846972b" containerName="extract" Nov 24 14:31:20 crc kubenswrapper[4822]: E1124 14:31:20.296314 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e971ff-ea2d-4652-901b-f744c846972b" containerName="pull" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.296322 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e971ff-ea2d-4652-901b-f744c846972b" containerName="pull" Nov 24 14:31:20 crc kubenswrapper[4822]: E1124 14:31:20.296336 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e971ff-ea2d-4652-901b-f744c846972b" containerName="util" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.296346 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e971ff-ea2d-4652-901b-f744c846972b" containerName="util" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.296469 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="20e971ff-ea2d-4652-901b-f744c846972b" containerName="extract" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.297528 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.301002 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.316963 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8"] Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.456603 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.456949 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.457026 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwc9r\" (UniqueName: \"kubernetes.io/projected/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-kube-api-access-cwc9r\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.558605 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwc9r\" (UniqueName: \"kubernetes.io/projected/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-kube-api-access-cwc9r\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.559014 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.559147 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.559719 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.559822 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.584160 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwc9r\" (UniqueName: \"kubernetes.io/projected/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-kube-api-access-cwc9r\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.626138 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:20 crc kubenswrapper[4822]: I1124 14:31:20.984719 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8"] Nov 24 14:31:21 crc kubenswrapper[4822]: I1124 14:31:21.451469 4822 generic.go:334] "Generic (PLEG): container finished" podID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerID="38c853ecc4af283b420bfc655314fe97118f216ea14016d1867cefff007d1093" exitCode=0 Nov 24 14:31:21 crc kubenswrapper[4822]: I1124 14:31:21.451523 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" event={"ID":"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0","Type":"ContainerDied","Data":"38c853ecc4af283b420bfc655314fe97118f216ea14016d1867cefff007d1093"} Nov 24 14:31:21 crc kubenswrapper[4822]: I1124 14:31:21.451556 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" event={"ID":"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0","Type":"ContainerStarted","Data":"fd14901ff75148a152c8f2abaa852681083b7a98179e733f2d3781330d8428a5"} Nov 24 14:31:24 crc kubenswrapper[4822]: I1124 14:31:24.475614 4822 generic.go:334] "Generic (PLEG): container finished" podID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerID="475cf02775078e75e61445c01ed1fa1d044f54b55005a3bf89828fb409a11baf" exitCode=0 Nov 24 14:31:24 crc kubenswrapper[4822]: I1124 14:31:24.475759 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" event={"ID":"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0","Type":"ContainerDied","Data":"475cf02775078e75e61445c01ed1fa1d044f54b55005a3bf89828fb409a11baf"} Nov 24 14:31:25 crc kubenswrapper[4822]: I1124 14:31:25.486611 4822 generic.go:334] "Generic (PLEG): container finished" podID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerID="1b928f73eed6404ef4dd994a7832bc017005286b4a4fe2eef2d35855c8ba557c" exitCode=0 Nov 24 14:31:25 crc kubenswrapper[4822]: I1124 14:31:25.486660 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" event={"ID":"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0","Type":"ContainerDied","Data":"1b928f73eed6404ef4dd994a7832bc017005286b4a4fe2eef2d35855c8ba557c"} Nov 24 14:31:26 crc kubenswrapper[4822]: I1124 14:31:26.810567 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:26 crc kubenswrapper[4822]: I1124 14:31:26.956274 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwc9r\" (UniqueName: \"kubernetes.io/projected/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-kube-api-access-cwc9r\") pod \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " Nov 24 14:31:26 crc kubenswrapper[4822]: I1124 14:31:26.956867 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-util\") pod \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " Nov 24 14:31:26 crc kubenswrapper[4822]: I1124 14:31:26.956978 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-bundle\") pod \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\" (UID: \"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0\") " Nov 24 14:31:26 crc kubenswrapper[4822]: I1124 14:31:26.957783 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-bundle" (OuterVolumeSpecName: "bundle") pod "1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" (UID: "1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:31:26 crc kubenswrapper[4822]: I1124 14:31:26.963379 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-kube-api-access-cwc9r" (OuterVolumeSpecName: "kube-api-access-cwc9r") pod "1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" (UID: "1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0"). InnerVolumeSpecName "kube-api-access-cwc9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:31:26 crc kubenswrapper[4822]: I1124 14:31:26.968269 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-util" (OuterVolumeSpecName: "util") pod "1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" (UID: "1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:31:27 crc kubenswrapper[4822]: I1124 14:31:27.058587 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwc9r\" (UniqueName: \"kubernetes.io/projected/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-kube-api-access-cwc9r\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:27 crc kubenswrapper[4822]: I1124 14:31:27.058641 4822 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-util\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:27 crc kubenswrapper[4822]: I1124 14:31:27.058664 4822 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:27 crc kubenswrapper[4822]: I1124 14:31:27.518679 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" event={"ID":"1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0","Type":"ContainerDied","Data":"fd14901ff75148a152c8f2abaa852681083b7a98179e733f2d3781330d8428a5"} Nov 24 14:31:27 crc kubenswrapper[4822]: I1124 14:31:27.518725 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd14901ff75148a152c8f2abaa852681083b7a98179e733f2d3781330d8428a5" Nov 24 14:31:27 crc kubenswrapper[4822]: I1124 14:31:27.518765 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.471429 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-hnmjh"] Nov 24 14:31:29 crc kubenswrapper[4822]: E1124 14:31:29.472049 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerName="pull" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.472072 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerName="pull" Nov 24 14:31:29 crc kubenswrapper[4822]: E1124 14:31:29.472098 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerName="util" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.472111 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerName="util" Nov 24 14:31:29 crc kubenswrapper[4822]: E1124 14:31:29.472135 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerName="extract" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.472151 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerName="extract" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.472343 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0" containerName="extract" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.472995 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-hnmjh" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.477563 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.477993 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.478283 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-hnmjh"] Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.479793 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-pv9pg" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.598243 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmn65\" (UniqueName: \"kubernetes.io/projected/08639b4b-1faa-4397-9727-b752d99978bc-kube-api-access-qmn65\") pod \"nmstate-operator-557fdffb88-hnmjh\" (UID: \"08639b4b-1faa-4397-9727-b752d99978bc\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-hnmjh" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.700759 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmn65\" (UniqueName: \"kubernetes.io/projected/08639b4b-1faa-4397-9727-b752d99978bc-kube-api-access-qmn65\") pod \"nmstate-operator-557fdffb88-hnmjh\" (UID: \"08639b4b-1faa-4397-9727-b752d99978bc\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-hnmjh" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.721701 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmn65\" (UniqueName: \"kubernetes.io/projected/08639b4b-1faa-4397-9727-b752d99978bc-kube-api-access-qmn65\") pod \"nmstate-operator-557fdffb88-hnmjh\" (UID: \"08639b4b-1faa-4397-9727-b752d99978bc\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-hnmjh" Nov 24 14:31:29 crc kubenswrapper[4822]: I1124 14:31:29.790725 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-hnmjh" Nov 24 14:31:30 crc kubenswrapper[4822]: I1124 14:31:30.243918 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-hnmjh"] Nov 24 14:31:30 crc kubenswrapper[4822]: I1124 14:31:30.540952 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-hnmjh" event={"ID":"08639b4b-1faa-4397-9727-b752d99978bc","Type":"ContainerStarted","Data":"5f10d4e40fe9a1046d02a418cdb6cc1049746a5bbcc636520551abfe30f6fef6"} Nov 24 14:31:32 crc kubenswrapper[4822]: I1124 14:31:32.558248 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-hnmjh" event={"ID":"08639b4b-1faa-4397-9727-b752d99978bc","Type":"ContainerStarted","Data":"84f87581a7a5f1b6514c9c988e816abc0a4f11308a8eaf9970d2bc288dd70958"} Nov 24 14:31:32 crc kubenswrapper[4822]: I1124 14:31:32.581677 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-hnmjh" podStartSLOduration=1.462964831 podStartE2EDuration="3.581649275s" podCreationTimestamp="2025-11-24 14:31:29 +0000 UTC" firstStartedPulling="2025-11-24 14:31:30.252871777 +0000 UTC m=+727.369512254" lastFinishedPulling="2025-11-24 14:31:32.371556211 +0000 UTC m=+729.488196698" observedRunningTime="2025-11-24 14:31:32.578632799 +0000 UTC m=+729.695273296" watchObservedRunningTime="2025-11-24 14:31:32.581649275 +0000 UTC m=+729.698289752" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.654368 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q"] Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.683316 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.691328 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-dwv7f" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.700714 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl"] Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.702730 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.720035 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.728447 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q"] Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.734754 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl"] Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.750847 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-xljmw"] Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.751921 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.762340 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4fcp\" (UniqueName: \"kubernetes.io/projected/56484cc3-efd5-4514-927c-350149fa1362-kube-api-access-n4fcp\") pod \"nmstate-webhook-6b89b748d8-zgvgl\" (UID: \"56484cc3-efd5-4514-927c-350149fa1362\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.762393 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp9x7\" (UniqueName: \"kubernetes.io/projected/1e2f143c-b018-4732-b91d-bfecee8726a9-kube-api-access-vp9x7\") pod \"nmstate-metrics-5dcf9c57c5-2ns9q\" (UID: \"1e2f143c-b018-4732-b91d-bfecee8726a9\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.762435 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/56484cc3-efd5-4514-927c-350149fa1362-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-zgvgl\" (UID: \"56484cc3-efd5-4514-927c-350149fa1362\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.827872 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb"] Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.828996 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.838310 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-mk5dj" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.838485 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.839065 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.842039 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb"] Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.865533 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp9x7\" (UniqueName: \"kubernetes.io/projected/1e2f143c-b018-4732-b91d-bfecee8726a9-kube-api-access-vp9x7\") pod \"nmstate-metrics-5dcf9c57c5-2ns9q\" (UID: \"1e2f143c-b018-4732-b91d-bfecee8726a9\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.865664 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/27e6c538-998e-46e2-acba-eb9eb622910f-ovs-socket\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.865794 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/56484cc3-efd5-4514-927c-350149fa1362-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-zgvgl\" (UID: \"56484cc3-efd5-4514-927c-350149fa1362\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.865864 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4pvm\" (UniqueName: \"kubernetes.io/projected/27e6c538-998e-46e2-acba-eb9eb622910f-kube-api-access-l4pvm\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.865915 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/27e6c538-998e-46e2-acba-eb9eb622910f-dbus-socket\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.865966 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/27e6c538-998e-46e2-acba-eb9eb622910f-nmstate-lock\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.866017 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4fcp\" (UniqueName: \"kubernetes.io/projected/56484cc3-efd5-4514-927c-350149fa1362-kube-api-access-n4fcp\") pod \"nmstate-webhook-6b89b748d8-zgvgl\" (UID: \"56484cc3-efd5-4514-927c-350149fa1362\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.880734 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/56484cc3-efd5-4514-927c-350149fa1362-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-zgvgl\" (UID: \"56484cc3-efd5-4514-927c-350149fa1362\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.885827 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4fcp\" (UniqueName: \"kubernetes.io/projected/56484cc3-efd5-4514-927c-350149fa1362-kube-api-access-n4fcp\") pod \"nmstate-webhook-6b89b748d8-zgvgl\" (UID: \"56484cc3-efd5-4514-927c-350149fa1362\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.886246 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp9x7\" (UniqueName: \"kubernetes.io/projected/1e2f143c-b018-4732-b91d-bfecee8726a9-kube-api-access-vp9x7\") pod \"nmstate-metrics-5dcf9c57c5-2ns9q\" (UID: \"1e2f143c-b018-4732-b91d-bfecee8726a9\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.966921 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhwf8\" (UniqueName: \"kubernetes.io/projected/4dc5455a-d1ab-40b6-942f-00e55730b89b-kube-api-access-jhwf8\") pod \"nmstate-console-plugin-5874bd7bc5-96lpb\" (UID: \"4dc5455a-d1ab-40b6-942f-00e55730b89b\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.966999 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dc5455a-d1ab-40b6-942f-00e55730b89b-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-96lpb\" (UID: \"4dc5455a-d1ab-40b6-942f-00e55730b89b\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.967151 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/27e6c538-998e-46e2-acba-eb9eb622910f-ovs-socket\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.967182 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4pvm\" (UniqueName: \"kubernetes.io/projected/27e6c538-998e-46e2-acba-eb9eb622910f-kube-api-access-l4pvm\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.967341 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/27e6c538-998e-46e2-acba-eb9eb622910f-ovs-socket\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.967688 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/27e6c538-998e-46e2-acba-eb9eb622910f-dbus-socket\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.967746 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4dc5455a-d1ab-40b6-942f-00e55730b89b-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-96lpb\" (UID: \"4dc5455a-d1ab-40b6-942f-00e55730b89b\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.967844 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/27e6c538-998e-46e2-acba-eb9eb622910f-nmstate-lock\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.967877 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/27e6c538-998e-46e2-acba-eb9eb622910f-dbus-socket\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:33 crc kubenswrapper[4822]: I1124 14:31:33.967965 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/27e6c538-998e-46e2-acba-eb9eb622910f-nmstate-lock\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.003649 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4pvm\" (UniqueName: \"kubernetes.io/projected/27e6c538-998e-46e2-acba-eb9eb622910f-kube-api-access-l4pvm\") pod \"nmstate-handler-xljmw\" (UID: \"27e6c538-998e-46e2-acba-eb9eb622910f\") " pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.021711 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.036083 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.060217 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-84fc68fcb9-h8m4z"] Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.062155 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.064751 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.071812 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4dc5455a-d1ab-40b6-942f-00e55730b89b-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-96lpb\" (UID: \"4dc5455a-d1ab-40b6-942f-00e55730b89b\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.071935 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhwf8\" (UniqueName: \"kubernetes.io/projected/4dc5455a-d1ab-40b6-942f-00e55730b89b-kube-api-access-jhwf8\") pod \"nmstate-console-plugin-5874bd7bc5-96lpb\" (UID: \"4dc5455a-d1ab-40b6-942f-00e55730b89b\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.071972 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dc5455a-d1ab-40b6-942f-00e55730b89b-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-96lpb\" (UID: \"4dc5455a-d1ab-40b6-942f-00e55730b89b\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.073587 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4dc5455a-d1ab-40b6-942f-00e55730b89b-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-96lpb\" (UID: \"4dc5455a-d1ab-40b6-942f-00e55730b89b\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.081004 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dc5455a-d1ab-40b6-942f-00e55730b89b-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-96lpb\" (UID: \"4dc5455a-d1ab-40b6-942f-00e55730b89b\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.082167 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-84fc68fcb9-h8m4z"] Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.099572 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhwf8\" (UniqueName: \"kubernetes.io/projected/4dc5455a-d1ab-40b6-942f-00e55730b89b-kube-api-access-jhwf8\") pod \"nmstate-console-plugin-5874bd7bc5-96lpb\" (UID: \"4dc5455a-d1ab-40b6-942f-00e55730b89b\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:34 crc kubenswrapper[4822]: W1124 14:31:34.105887 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27e6c538_998e_46e2_acba_eb9eb622910f.slice/crio-d3be6a81648a52b6d152c63cf19db1a42345e6323425c301f7d6d22ecbe12658 WatchSource:0}: Error finding container d3be6a81648a52b6d152c63cf19db1a42345e6323425c301f7d6d22ecbe12658: Status 404 returned error can't find the container with id d3be6a81648a52b6d152c63cf19db1a42345e6323425c301f7d6d22ecbe12658 Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.173607 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-service-ca\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.174050 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-console-config\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.174225 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj2zv\" (UniqueName: \"kubernetes.io/projected/7dd32891-20d0-477c-8863-329c22e547ae-kube-api-access-vj2zv\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.175042 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-oauth-serving-cert\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.175163 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-trusted-ca-bundle\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.175380 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7dd32891-20d0-477c-8863-329c22e547ae-console-oauth-config\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.175475 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7dd32891-20d0-477c-8863-329c22e547ae-console-serving-cert\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.226696 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.276535 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-service-ca\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.276596 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-console-config\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.276652 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj2zv\" (UniqueName: \"kubernetes.io/projected/7dd32891-20d0-477c-8863-329c22e547ae-kube-api-access-vj2zv\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.276689 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-trusted-ca-bundle\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.276710 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-oauth-serving-cert\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.276962 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7dd32891-20d0-477c-8863-329c22e547ae-console-oauth-config\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.277569 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7dd32891-20d0-477c-8863-329c22e547ae-console-serving-cert\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.279960 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-console-config\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.280457 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-oauth-serving-cert\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.281676 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-service-ca\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.287245 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dd32891-20d0-477c-8863-329c22e547ae-trusted-ca-bundle\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.288366 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7dd32891-20d0-477c-8863-329c22e547ae-console-oauth-config\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.288870 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7dd32891-20d0-477c-8863-329c22e547ae-console-serving-cert\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.294426 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj2zv\" (UniqueName: \"kubernetes.io/projected/7dd32891-20d0-477c-8863-329c22e547ae-kube-api-access-vj2zv\") pod \"console-84fc68fcb9-h8m4z\" (UID: \"7dd32891-20d0-477c-8863-329c22e547ae\") " pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.421812 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.484164 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb"] Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.546979 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl"] Nov 24 14:31:34 crc kubenswrapper[4822]: W1124 14:31:34.551459 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56484cc3_efd5_4514_927c_350149fa1362.slice/crio-e68004608c0ff056ff8174e52593227d382cf6fdff608eba71dbabbe081fac0a WatchSource:0}: Error finding container e68004608c0ff056ff8174e52593227d382cf6fdff608eba71dbabbe081fac0a: Status 404 returned error can't find the container with id e68004608c0ff056ff8174e52593227d382cf6fdff608eba71dbabbe081fac0a Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.570685 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" event={"ID":"56484cc3-efd5-4514-927c-350149fa1362","Type":"ContainerStarted","Data":"e68004608c0ff056ff8174e52593227d382cf6fdff608eba71dbabbe081fac0a"} Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.572259 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xljmw" event={"ID":"27e6c538-998e-46e2-acba-eb9eb622910f","Type":"ContainerStarted","Data":"d3be6a81648a52b6d152c63cf19db1a42345e6323425c301f7d6d22ecbe12658"} Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.573509 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" event={"ID":"4dc5455a-d1ab-40b6-942f-00e55730b89b","Type":"ContainerStarted","Data":"2d3daae3fc00e22eb7588bbfa06c46d992d850de7d66289ed9f608045971785c"} Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.626468 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q"] Nov 24 14:31:34 crc kubenswrapper[4822]: W1124 14:31:34.637674 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e2f143c_b018_4732_b91d_bfecee8726a9.slice/crio-20b0445caede45f9fb30de24a9bf304f5647642c75dfe159ea0eae5ea6073f8d WatchSource:0}: Error finding container 20b0445caede45f9fb30de24a9bf304f5647642c75dfe159ea0eae5ea6073f8d: Status 404 returned error can't find the container with id 20b0445caede45f9fb30de24a9bf304f5647642c75dfe159ea0eae5ea6073f8d Nov 24 14:31:34 crc kubenswrapper[4822]: I1124 14:31:34.720005 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-84fc68fcb9-h8m4z"] Nov 24 14:31:34 crc kubenswrapper[4822]: W1124 14:31:34.720252 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dd32891_20d0_477c_8863_329c22e547ae.slice/crio-a85ff0a00a3f055fb181b5c52a8dfc3dfc1bfe5c9597a7ebbb8f4bbd5d8b7132 WatchSource:0}: Error finding container a85ff0a00a3f055fb181b5c52a8dfc3dfc1bfe5c9597a7ebbb8f4bbd5d8b7132: Status 404 returned error can't find the container with id a85ff0a00a3f055fb181b5c52a8dfc3dfc1bfe5c9597a7ebbb8f4bbd5d8b7132 Nov 24 14:31:35 crc kubenswrapper[4822]: I1124 14:31:35.580076 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q" event={"ID":"1e2f143c-b018-4732-b91d-bfecee8726a9","Type":"ContainerStarted","Data":"20b0445caede45f9fb30de24a9bf304f5647642c75dfe159ea0eae5ea6073f8d"} Nov 24 14:31:35 crc kubenswrapper[4822]: I1124 14:31:35.582115 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-84fc68fcb9-h8m4z" event={"ID":"7dd32891-20d0-477c-8863-329c22e547ae","Type":"ContainerStarted","Data":"08741f3a00b9e71b385854bbf811e73454988e71fa1bccd8fe6024fab686abd1"} Nov 24 14:31:35 crc kubenswrapper[4822]: I1124 14:31:35.582145 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-84fc68fcb9-h8m4z" event={"ID":"7dd32891-20d0-477c-8863-329c22e547ae","Type":"ContainerStarted","Data":"a85ff0a00a3f055fb181b5c52a8dfc3dfc1bfe5c9597a7ebbb8f4bbd5d8b7132"} Nov 24 14:31:35 crc kubenswrapper[4822]: I1124 14:31:35.612094 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-84fc68fcb9-h8m4z" podStartSLOduration=1.612061795 podStartE2EDuration="1.612061795s" podCreationTimestamp="2025-11-24 14:31:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:31:35.60503868 +0000 UTC m=+732.721679157" watchObservedRunningTime="2025-11-24 14:31:35.612061795 +0000 UTC m=+732.728702302" Nov 24 14:31:37 crc kubenswrapper[4822]: I1124 14:31:37.601448 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q" event={"ID":"1e2f143c-b018-4732-b91d-bfecee8726a9","Type":"ContainerStarted","Data":"8ff1a3ffa2d197e850b5735712358cae51b98e601e414ee31df0cad5a156817c"} Nov 24 14:31:37 crc kubenswrapper[4822]: I1124 14:31:37.606055 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xljmw" event={"ID":"27e6c538-998e-46e2-acba-eb9eb622910f","Type":"ContainerStarted","Data":"bd52b2efb00b1f5b674b7d7160a341e3658bc258b70b51aaf5e074b8ea2a3edf"} Nov 24 14:31:37 crc kubenswrapper[4822]: I1124 14:31:37.606383 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:37 crc kubenswrapper[4822]: I1124 14:31:37.630514 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-xljmw" podStartSLOduration=1.913503443 podStartE2EDuration="4.630495266s" podCreationTimestamp="2025-11-24 14:31:33 +0000 UTC" firstStartedPulling="2025-11-24 14:31:34.108058023 +0000 UTC m=+731.224698500" lastFinishedPulling="2025-11-24 14:31:36.825049846 +0000 UTC m=+733.941690323" observedRunningTime="2025-11-24 14:31:37.627128428 +0000 UTC m=+734.743768905" watchObservedRunningTime="2025-11-24 14:31:37.630495266 +0000 UTC m=+734.747135743" Nov 24 14:31:38 crc kubenswrapper[4822]: I1124 14:31:38.616176 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" event={"ID":"4dc5455a-d1ab-40b6-942f-00e55730b89b","Type":"ContainerStarted","Data":"e46697867cdcc11e625e16a07c004cce6841826f30aa01ac45e18a2534e13c16"} Nov 24 14:31:38 crc kubenswrapper[4822]: I1124 14:31:38.637610 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-96lpb" podStartSLOduration=2.069132974 podStartE2EDuration="5.637576227s" podCreationTimestamp="2025-11-24 14:31:33 +0000 UTC" firstStartedPulling="2025-11-24 14:31:34.495579136 +0000 UTC m=+731.612219613" lastFinishedPulling="2025-11-24 14:31:38.064022389 +0000 UTC m=+735.180662866" observedRunningTime="2025-11-24 14:31:38.635396708 +0000 UTC m=+735.752037195" watchObservedRunningTime="2025-11-24 14:31:38.637576227 +0000 UTC m=+735.754216734" Nov 24 14:31:40 crc kubenswrapper[4822]: I1124 14:31:40.657504 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" event={"ID":"56484cc3-efd5-4514-927c-350149fa1362","Type":"ContainerStarted","Data":"af72a6472f77272bc896c8e9d6d8ee9d3ec87ac1769484f85dcf6b3cbdb61cc2"} Nov 24 14:31:40 crc kubenswrapper[4822]: I1124 14:31:40.658738 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:31:40 crc kubenswrapper[4822]: I1124 14:31:40.689453 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q" event={"ID":"1e2f143c-b018-4732-b91d-bfecee8726a9","Type":"ContainerStarted","Data":"d0c8de26eb08c5c69dab995d3e8acab0bec09a4753bad57aa2836a7a36076dbe"} Nov 24 14:31:40 crc kubenswrapper[4822]: I1124 14:31:40.726087 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2ns9q" podStartSLOduration=2.685837023 podStartE2EDuration="7.726067187s" podCreationTimestamp="2025-11-24 14:31:33 +0000 UTC" firstStartedPulling="2025-11-24 14:31:34.640050403 +0000 UTC m=+731.756690880" lastFinishedPulling="2025-11-24 14:31:39.680280567 +0000 UTC m=+736.796921044" observedRunningTime="2025-11-24 14:31:40.724527287 +0000 UTC m=+737.841167764" watchObservedRunningTime="2025-11-24 14:31:40.726067187 +0000 UTC m=+737.842707654" Nov 24 14:31:40 crc kubenswrapper[4822]: I1124 14:31:40.727452 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" podStartSLOduration=2.574829915 podStartE2EDuration="7.72744622s" podCreationTimestamp="2025-11-24 14:31:33 +0000 UTC" firstStartedPulling="2025-11-24 14:31:34.553355953 +0000 UTC m=+731.669996430" lastFinishedPulling="2025-11-24 14:31:39.705972258 +0000 UTC m=+736.822612735" observedRunningTime="2025-11-24 14:31:40.694740465 +0000 UTC m=+737.811380952" watchObservedRunningTime="2025-11-24 14:31:40.72744622 +0000 UTC m=+737.844086697" Nov 24 14:31:41 crc kubenswrapper[4822]: I1124 14:31:41.578130 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:31:41 crc kubenswrapper[4822]: I1124 14:31:41.578249 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:31:44 crc kubenswrapper[4822]: I1124 14:31:44.104819 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-xljmw" Nov 24 14:31:44 crc kubenswrapper[4822]: I1124 14:31:44.422716 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:44 crc kubenswrapper[4822]: I1124 14:31:44.422794 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:44 crc kubenswrapper[4822]: I1124 14:31:44.431370 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:44 crc kubenswrapper[4822]: I1124 14:31:44.724282 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-84fc68fcb9-h8m4z" Nov 24 14:31:44 crc kubenswrapper[4822]: I1124 14:31:44.805632 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5v7wp"] Nov 24 14:31:54 crc kubenswrapper[4822]: I1124 14:31:54.045607 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-zgvgl" Nov 24 14:32:00 crc kubenswrapper[4822]: I1124 14:32:00.753952 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hl4bh"] Nov 24 14:32:00 crc kubenswrapper[4822]: I1124 14:32:00.754806 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" podUID="72315396-ab30-4736-8eb9-f735ef0e7f97" containerName="controller-manager" containerID="cri-o://13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f" gracePeriod=30 Nov 24 14:32:00 crc kubenswrapper[4822]: I1124 14:32:00.849605 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8"] Nov 24 14:32:00 crc kubenswrapper[4822]: I1124 14:32:00.850092 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" podUID="397bb803-fafb-4fca-ae9f-e373d413958f" containerName="route-controller-manager" containerID="cri-o://65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc" gracePeriod=30 Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.367002 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.410178 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-client-ca\") pod \"72315396-ab30-4736-8eb9-f735ef0e7f97\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.410268 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-config\") pod \"72315396-ab30-4736-8eb9-f735ef0e7f97\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.410314 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-proxy-ca-bundles\") pod \"72315396-ab30-4736-8eb9-f735ef0e7f97\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.410380 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72315396-ab30-4736-8eb9-f735ef0e7f97-serving-cert\") pod \"72315396-ab30-4736-8eb9-f735ef0e7f97\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.410410 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkk4s\" (UniqueName: \"kubernetes.io/projected/72315396-ab30-4736-8eb9-f735ef0e7f97-kube-api-access-hkk4s\") pod \"72315396-ab30-4736-8eb9-f735ef0e7f97\" (UID: \"72315396-ab30-4736-8eb9-f735ef0e7f97\") " Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.411328 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-client-ca" (OuterVolumeSpecName: "client-ca") pod "72315396-ab30-4736-8eb9-f735ef0e7f97" (UID: "72315396-ab30-4736-8eb9-f735ef0e7f97"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.411649 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "72315396-ab30-4736-8eb9-f735ef0e7f97" (UID: "72315396-ab30-4736-8eb9-f735ef0e7f97"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.416045 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-config" (OuterVolumeSpecName: "config") pod "72315396-ab30-4736-8eb9-f735ef0e7f97" (UID: "72315396-ab30-4736-8eb9-f735ef0e7f97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.418111 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72315396-ab30-4736-8eb9-f735ef0e7f97-kube-api-access-hkk4s" (OuterVolumeSpecName: "kube-api-access-hkk4s") pod "72315396-ab30-4736-8eb9-f735ef0e7f97" (UID: "72315396-ab30-4736-8eb9-f735ef0e7f97"). InnerVolumeSpecName "kube-api-access-hkk4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.420923 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.430064 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72315396-ab30-4736-8eb9-f735ef0e7f97-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "72315396-ab30-4736-8eb9-f735ef0e7f97" (UID: "72315396-ab30-4736-8eb9-f735ef0e7f97"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.512250 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-config\") pod \"397bb803-fafb-4fca-ae9f-e373d413958f\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.512423 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-client-ca\") pod \"397bb803-fafb-4fca-ae9f-e373d413958f\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.512487 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68k2q\" (UniqueName: \"kubernetes.io/projected/397bb803-fafb-4fca-ae9f-e373d413958f-kube-api-access-68k2q\") pod \"397bb803-fafb-4fca-ae9f-e373d413958f\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.512554 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/397bb803-fafb-4fca-ae9f-e373d413958f-serving-cert\") pod \"397bb803-fafb-4fca-ae9f-e373d413958f\" (UID: \"397bb803-fafb-4fca-ae9f-e373d413958f\") " Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.512926 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72315396-ab30-4736-8eb9-f735ef0e7f97-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.512951 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkk4s\" (UniqueName: \"kubernetes.io/projected/72315396-ab30-4736-8eb9-f735ef0e7f97-kube-api-access-hkk4s\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.512965 4822 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.512978 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.512991 4822 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72315396-ab30-4736-8eb9-f735ef0e7f97-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.513247 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-client-ca" (OuterVolumeSpecName: "client-ca") pod "397bb803-fafb-4fca-ae9f-e373d413958f" (UID: "397bb803-fafb-4fca-ae9f-e373d413958f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.513293 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-config" (OuterVolumeSpecName: "config") pod "397bb803-fafb-4fca-ae9f-e373d413958f" (UID: "397bb803-fafb-4fca-ae9f-e373d413958f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.519802 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/397bb803-fafb-4fca-ae9f-e373d413958f-kube-api-access-68k2q" (OuterVolumeSpecName: "kube-api-access-68k2q") pod "397bb803-fafb-4fca-ae9f-e373d413958f" (UID: "397bb803-fafb-4fca-ae9f-e373d413958f"). InnerVolumeSpecName "kube-api-access-68k2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.523351 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/397bb803-fafb-4fca-ae9f-e373d413958f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "397bb803-fafb-4fca-ae9f-e373d413958f" (UID: "397bb803-fafb-4fca-ae9f-e373d413958f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.614495 4822 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.614532 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68k2q\" (UniqueName: \"kubernetes.io/projected/397bb803-fafb-4fca-ae9f-e373d413958f-kube-api-access-68k2q\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.614547 4822 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/397bb803-fafb-4fca-ae9f-e373d413958f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.614559 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/397bb803-fafb-4fca-ae9f-e373d413958f-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.860511 4822 generic.go:334] "Generic (PLEG): container finished" podID="397bb803-fafb-4fca-ae9f-e373d413958f" containerID="65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc" exitCode=0 Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.860581 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" event={"ID":"397bb803-fafb-4fca-ae9f-e373d413958f","Type":"ContainerDied","Data":"65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc"} Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.860615 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" event={"ID":"397bb803-fafb-4fca-ae9f-e373d413958f","Type":"ContainerDied","Data":"0b83989637e7062aab70e259f63105b0a72baf106f0415824ce1b59d3e404da3"} Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.860636 4822 scope.go:117] "RemoveContainer" containerID="65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.860751 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.876933 4822 generic.go:334] "Generic (PLEG): container finished" podID="72315396-ab30-4736-8eb9-f735ef0e7f97" containerID="13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f" exitCode=0 Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.877488 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" event={"ID":"72315396-ab30-4736-8eb9-f735ef0e7f97","Type":"ContainerDied","Data":"13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f"} Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.877659 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" event={"ID":"72315396-ab30-4736-8eb9-f735ef0e7f97","Type":"ContainerDied","Data":"3a9098f789c9bc1afb342643ff741d594840f520c39b017051d9b747fd8dbcdb"} Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.877799 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hl4bh" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.890355 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8"] Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.894680 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7wf8"] Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.910666 4822 scope.go:117] "RemoveContainer" containerID="65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc" Nov 24 14:32:01 crc kubenswrapper[4822]: E1124 14:32:01.911362 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc\": container with ID starting with 65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc not found: ID does not exist" containerID="65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.911396 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc"} err="failed to get container status \"65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc\": rpc error: code = NotFound desc = could not find container \"65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc\": container with ID starting with 65767a97518e95c57059f50b103be56a2a0a419b6d6ce60d0091fcec438385cc not found: ID does not exist" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.911424 4822 scope.go:117] "RemoveContainer" containerID="13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.924830 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hl4bh"] Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.928331 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hl4bh"] Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.946863 4822 scope.go:117] "RemoveContainer" containerID="13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f" Nov 24 14:32:01 crc kubenswrapper[4822]: E1124 14:32:01.947501 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f\": container with ID starting with 13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f not found: ID does not exist" containerID="13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f" Nov 24 14:32:01 crc kubenswrapper[4822]: I1124 14:32:01.947606 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f"} err="failed to get container status \"13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f\": rpc error: code = NotFound desc = could not find container \"13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f\": container with ID starting with 13a5dd0b599da875c8119eedf6edad5625265eaf012b1759f6b70d92cc66883f not found: ID does not exist" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.734966 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb"] Nov 24 14:32:02 crc kubenswrapper[4822]: E1124 14:32:02.735656 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72315396-ab30-4736-8eb9-f735ef0e7f97" containerName="controller-manager" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.735671 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="72315396-ab30-4736-8eb9-f735ef0e7f97" containerName="controller-manager" Nov 24 14:32:02 crc kubenswrapper[4822]: E1124 14:32:02.735703 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="397bb803-fafb-4fca-ae9f-e373d413958f" containerName="route-controller-manager" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.735711 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="397bb803-fafb-4fca-ae9f-e373d413958f" containerName="route-controller-manager" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.735857 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="72315396-ab30-4736-8eb9-f735ef0e7f97" containerName="controller-manager" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.735875 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="397bb803-fafb-4fca-ae9f-e373d413958f" containerName="route-controller-manager" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.736478 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.740997 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.741831 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.742222 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.742510 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.743078 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.743753 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.751193 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-ffff65865-j56xj"] Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.754429 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.755417 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-ffff65865-j56xj"] Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.762953 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.763007 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.763362 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.763522 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.763855 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.764070 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.770901 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb"] Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.772777 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.845358 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f70218a5-716e-4451-8ef0-487f1744dca0-config\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.845409 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p4dr\" (UniqueName: \"kubernetes.io/projected/618c9342-120e-457d-af0d-e2bdf42ed450-kube-api-access-5p4dr\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.845441 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/618c9342-120e-457d-af0d-e2bdf42ed450-client-ca\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.845460 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/618c9342-120e-457d-af0d-e2bdf42ed450-serving-cert\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.845488 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2vhp\" (UniqueName: \"kubernetes.io/projected/f70218a5-716e-4451-8ef0-487f1744dca0-kube-api-access-v2vhp\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.845511 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f70218a5-716e-4451-8ef0-487f1744dca0-client-ca\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.845531 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f70218a5-716e-4451-8ef0-487f1744dca0-serving-cert\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.845554 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618c9342-120e-457d-af0d-e2bdf42ed450-config\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.845598 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/618c9342-120e-457d-af0d-e2bdf42ed450-proxy-ca-bundles\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.947012 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2vhp\" (UniqueName: \"kubernetes.io/projected/f70218a5-716e-4451-8ef0-487f1744dca0-kube-api-access-v2vhp\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.947058 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f70218a5-716e-4451-8ef0-487f1744dca0-client-ca\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.947083 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f70218a5-716e-4451-8ef0-487f1744dca0-serving-cert\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.947112 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618c9342-120e-457d-af0d-e2bdf42ed450-config\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.947168 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/618c9342-120e-457d-af0d-e2bdf42ed450-proxy-ca-bundles\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.947232 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f70218a5-716e-4451-8ef0-487f1744dca0-config\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.947261 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p4dr\" (UniqueName: \"kubernetes.io/projected/618c9342-120e-457d-af0d-e2bdf42ed450-kube-api-access-5p4dr\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.947286 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/618c9342-120e-457d-af0d-e2bdf42ed450-serving-cert\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.947308 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/618c9342-120e-457d-af0d-e2bdf42ed450-client-ca\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.948422 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/618c9342-120e-457d-af0d-e2bdf42ed450-client-ca\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.949276 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/618c9342-120e-457d-af0d-e2bdf42ed450-proxy-ca-bundles\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.950053 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f70218a5-716e-4451-8ef0-487f1744dca0-client-ca\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.951497 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f70218a5-716e-4451-8ef0-487f1744dca0-config\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.951952 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618c9342-120e-457d-af0d-e2bdf42ed450-config\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.958425 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f70218a5-716e-4451-8ef0-487f1744dca0-serving-cert\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.960085 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/618c9342-120e-457d-af0d-e2bdf42ed450-serving-cert\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.980624 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p4dr\" (UniqueName: \"kubernetes.io/projected/618c9342-120e-457d-af0d-e2bdf42ed450-kube-api-access-5p4dr\") pod \"controller-manager-ffff65865-j56xj\" (UID: \"618c9342-120e-457d-af0d-e2bdf42ed450\") " pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:02 crc kubenswrapper[4822]: I1124 14:32:02.988755 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2vhp\" (UniqueName: \"kubernetes.io/projected/f70218a5-716e-4451-8ef0-487f1744dca0-kube-api-access-v2vhp\") pod \"route-controller-manager-664f8874b9-2ftzb\" (UID: \"f70218a5-716e-4451-8ef0-487f1744dca0\") " pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.067741 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.094296 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.341524 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb"] Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.394270 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-ffff65865-j56xj"] Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.717434 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="397bb803-fafb-4fca-ae9f-e373d413958f" path="/var/lib/kubelet/pods/397bb803-fafb-4fca-ae9f-e373d413958f/volumes" Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.719079 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72315396-ab30-4736-8eb9-f735ef0e7f97" path="/var/lib/kubelet/pods/72315396-ab30-4736-8eb9-f735ef0e7f97/volumes" Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.907885 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" event={"ID":"f70218a5-716e-4451-8ef0-487f1744dca0","Type":"ContainerStarted","Data":"6c666a56644006590377f2a8dff93305de5b33e667c92612872cb3715e3b84b4"} Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.908380 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" event={"ID":"f70218a5-716e-4451-8ef0-487f1744dca0","Type":"ContainerStarted","Data":"78bec2304f41dfb785630c6961ed8525eb0eecddebce6cd8fb84e3edc9f842aa"} Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.908400 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.910129 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" event={"ID":"618c9342-120e-457d-af0d-e2bdf42ed450","Type":"ContainerStarted","Data":"86f082001fd2965cd78040e439426ba66e476faeb8c61a834f8e2f374ee97b1d"} Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.910177 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" event={"ID":"618c9342-120e-457d-af0d-e2bdf42ed450","Type":"ContainerStarted","Data":"2aa33c4a3f4e7e3d08aa58cb99bf9d432373d9fd16900794b59e8833bab490e9"} Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.910436 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.931449 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.934513 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" podStartSLOduration=3.934488112 podStartE2EDuration="3.934488112s" podCreationTimestamp="2025-11-24 14:32:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:32:03.930096071 +0000 UTC m=+761.046736568" watchObservedRunningTime="2025-11-24 14:32:03.934488112 +0000 UTC m=+761.051128589" Nov 24 14:32:03 crc kubenswrapper[4822]: I1124 14:32:03.956559 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-ffff65865-j56xj" podStartSLOduration=3.956536397 podStartE2EDuration="3.956536397s" podCreationTimestamp="2025-11-24 14:32:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:32:03.950712651 +0000 UTC m=+761.067353148" watchObservedRunningTime="2025-11-24 14:32:03.956536397 +0000 UTC m=+761.073176874" Nov 24 14:32:04 crc kubenswrapper[4822]: I1124 14:32:04.186961 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-664f8874b9-2ftzb" Nov 24 14:32:08 crc kubenswrapper[4822]: I1124 14:32:08.838113 4822 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 14:32:09 crc kubenswrapper[4822]: I1124 14:32:09.870439 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5v7wp" podUID="6153af6d-c527-485f-8a0e-0cfecfb43d5b" containerName="console" containerID="cri-o://b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656" gracePeriod=15 Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.413546 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5v7wp_6153af6d-c527-485f-8a0e-0cfecfb43d5b/console/0.log" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.414031 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.493628 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-config\") pod \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.493713 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-serving-cert\") pod \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.493774 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-oauth-serving-cert\") pod \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.493835 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dvgh\" (UniqueName: \"kubernetes.io/projected/6153af6d-c527-485f-8a0e-0cfecfb43d5b-kube-api-access-6dvgh\") pod \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.493950 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-service-ca\") pod \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.494015 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-oauth-config\") pod \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.494869 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6153af6d-c527-485f-8a0e-0cfecfb43d5b" (UID: "6153af6d-c527-485f-8a0e-0cfecfb43d5b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.494914 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-config" (OuterVolumeSpecName: "console-config") pod "6153af6d-c527-485f-8a0e-0cfecfb43d5b" (UID: "6153af6d-c527-485f-8a0e-0cfecfb43d5b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.494957 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-service-ca" (OuterVolumeSpecName: "service-ca") pod "6153af6d-c527-485f-8a0e-0cfecfb43d5b" (UID: "6153af6d-c527-485f-8a0e-0cfecfb43d5b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.495066 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6153af6d-c527-485f-8a0e-0cfecfb43d5b" (UID: "6153af6d-c527-485f-8a0e-0cfecfb43d5b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.495249 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-trusted-ca-bundle\") pod \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\" (UID: \"6153af6d-c527-485f-8a0e-0cfecfb43d5b\") " Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.495632 4822 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.495651 4822 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.495665 4822 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.495678 4822 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6153af6d-c527-485f-8a0e-0cfecfb43d5b-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.501334 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6153af6d-c527-485f-8a0e-0cfecfb43d5b-kube-api-access-6dvgh" (OuterVolumeSpecName: "kube-api-access-6dvgh") pod "6153af6d-c527-485f-8a0e-0cfecfb43d5b" (UID: "6153af6d-c527-485f-8a0e-0cfecfb43d5b"). InnerVolumeSpecName "kube-api-access-6dvgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.502183 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6153af6d-c527-485f-8a0e-0cfecfb43d5b" (UID: "6153af6d-c527-485f-8a0e-0cfecfb43d5b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.502812 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6153af6d-c527-485f-8a0e-0cfecfb43d5b" (UID: "6153af6d-c527-485f-8a0e-0cfecfb43d5b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.597757 4822 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.597806 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dvgh\" (UniqueName: \"kubernetes.io/projected/6153af6d-c527-485f-8a0e-0cfecfb43d5b-kube-api-access-6dvgh\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.597817 4822 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6153af6d-c527-485f-8a0e-0cfecfb43d5b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.980501 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5v7wp_6153af6d-c527-485f-8a0e-0cfecfb43d5b/console/0.log" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.981050 4822 generic.go:334] "Generic (PLEG): container finished" podID="6153af6d-c527-485f-8a0e-0cfecfb43d5b" containerID="b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656" exitCode=2 Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.981105 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5v7wp" event={"ID":"6153af6d-c527-485f-8a0e-0cfecfb43d5b","Type":"ContainerDied","Data":"b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656"} Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.981148 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5v7wp" event={"ID":"6153af6d-c527-485f-8a0e-0cfecfb43d5b","Type":"ContainerDied","Data":"9630f913db10497b21d6a0ef428e8ca730501d9c2a427ec53e0e1299d4ef37a5"} Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.981152 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5v7wp" Nov 24 14:32:10 crc kubenswrapper[4822]: I1124 14:32:10.981172 4822 scope.go:117] "RemoveContainer" containerID="b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656" Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.007982 4822 scope.go:117] "RemoveContainer" containerID="b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656" Nov 24 14:32:11 crc kubenswrapper[4822]: E1124 14:32:11.010821 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656\": container with ID starting with b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656 not found: ID does not exist" containerID="b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656" Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.010883 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656"} err="failed to get container status \"b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656\": rpc error: code = NotFound desc = could not find container \"b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656\": container with ID starting with b5f286d9af726d983fb3a130e2cdbda55d078c2bc4935758dc8a18af98d2e656 not found: ID does not exist" Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.034143 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5v7wp"] Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.043550 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5v7wp"] Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.578242 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.578314 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.578365 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.579266 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"065577b1f1af0833feb9bb8a7aed6dd3738bba738b7c830369d5c019d2095f70"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.579398 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://065577b1f1af0833feb9bb8a7aed6dd3738bba738b7c830369d5c019d2095f70" gracePeriod=600 Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.713440 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6153af6d-c527-485f-8a0e-0cfecfb43d5b" path="/var/lib/kubelet/pods/6153af6d-c527-485f-8a0e-0cfecfb43d5b/volumes" Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.992424 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="065577b1f1af0833feb9bb8a7aed6dd3738bba738b7c830369d5c019d2095f70" exitCode=0 Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.992492 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"065577b1f1af0833feb9bb8a7aed6dd3738bba738b7c830369d5c019d2095f70"} Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.993995 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"cb8615da8ae4badf56e6f90df575844458ac64f06801fa8ab7085476265bc5df"} Nov 24 14:32:11 crc kubenswrapper[4822]: I1124 14:32:11.994087 4822 scope.go:117] "RemoveContainer" containerID="c48afa8921812e0254372f9525778df4abed8a487a31452c711ba4cfd0103ae3" Nov 24 14:32:12 crc kubenswrapper[4822]: I1124 14:32:12.910792 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94"] Nov 24 14:32:12 crc kubenswrapper[4822]: E1124 14:32:12.911179 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6153af6d-c527-485f-8a0e-0cfecfb43d5b" containerName="console" Nov 24 14:32:12 crc kubenswrapper[4822]: I1124 14:32:12.911208 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="6153af6d-c527-485f-8a0e-0cfecfb43d5b" containerName="console" Nov 24 14:32:12 crc kubenswrapper[4822]: I1124 14:32:12.911407 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="6153af6d-c527-485f-8a0e-0cfecfb43d5b" containerName="console" Nov 24 14:32:12 crc kubenswrapper[4822]: I1124 14:32:12.912781 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:12 crc kubenswrapper[4822]: I1124 14:32:12.916698 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 14:32:12 crc kubenswrapper[4822]: I1124 14:32:12.927422 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94"] Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.043414 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.044010 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.044080 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqsf2\" (UniqueName: \"kubernetes.io/projected/cacd21c8-e424-492a-ba96-b618800b3d58-kube-api-access-vqsf2\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.146101 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.146173 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.146319 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqsf2\" (UniqueName: \"kubernetes.io/projected/cacd21c8-e424-492a-ba96-b618800b3d58-kube-api-access-vqsf2\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.146876 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.146953 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.167349 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqsf2\" (UniqueName: \"kubernetes.io/projected/cacd21c8-e424-492a-ba96-b618800b3d58-kube-api-access-vqsf2\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.233306 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:13 crc kubenswrapper[4822]: I1124 14:32:13.874004 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94"] Nov 24 14:32:13 crc kubenswrapper[4822]: W1124 14:32:13.882235 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcacd21c8_e424_492a_ba96_b618800b3d58.slice/crio-6e0fb67d7225c3911beaefcbcbd7304852e50f6f532d9949fcd20fb581d301bb WatchSource:0}: Error finding container 6e0fb67d7225c3911beaefcbcbd7304852e50f6f532d9949fcd20fb581d301bb: Status 404 returned error can't find the container with id 6e0fb67d7225c3911beaefcbcbd7304852e50f6f532d9949fcd20fb581d301bb Nov 24 14:32:14 crc kubenswrapper[4822]: I1124 14:32:14.010981 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" event={"ID":"cacd21c8-e424-492a-ba96-b618800b3d58","Type":"ContainerStarted","Data":"6e0fb67d7225c3911beaefcbcbd7304852e50f6f532d9949fcd20fb581d301bb"} Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.018769 4822 generic.go:334] "Generic (PLEG): container finished" podID="cacd21c8-e424-492a-ba96-b618800b3d58" containerID="9c9ba0949bc383731b8907ebeaade18f2541df4dd0b6024006cbc88b67db7974" exitCode=0 Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.018843 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" event={"ID":"cacd21c8-e424-492a-ba96-b618800b3d58","Type":"ContainerDied","Data":"9c9ba0949bc383731b8907ebeaade18f2541df4dd0b6024006cbc88b67db7974"} Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.255385 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6hldt"] Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.258026 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.294240 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6hldt"] Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.382346 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-catalog-content\") pod \"redhat-operators-6hldt\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.382598 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs6cg\" (UniqueName: \"kubernetes.io/projected/8169b355-d354-4eb8-8abc-fe1cc44b8557-kube-api-access-rs6cg\") pod \"redhat-operators-6hldt\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.382667 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-utilities\") pod \"redhat-operators-6hldt\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.484699 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-catalog-content\") pod \"redhat-operators-6hldt\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.484774 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs6cg\" (UniqueName: \"kubernetes.io/projected/8169b355-d354-4eb8-8abc-fe1cc44b8557-kube-api-access-rs6cg\") pod \"redhat-operators-6hldt\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.484809 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-utilities\") pod \"redhat-operators-6hldt\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.485357 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-catalog-content\") pod \"redhat-operators-6hldt\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.485482 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-utilities\") pod \"redhat-operators-6hldt\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.513302 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs6cg\" (UniqueName: \"kubernetes.io/projected/8169b355-d354-4eb8-8abc-fe1cc44b8557-kube-api-access-rs6cg\") pod \"redhat-operators-6hldt\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:15 crc kubenswrapper[4822]: I1124 14:32:15.601986 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:16 crc kubenswrapper[4822]: I1124 14:32:16.121054 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6hldt"] Nov 24 14:32:17 crc kubenswrapper[4822]: I1124 14:32:17.034609 4822 generic.go:334] "Generic (PLEG): container finished" podID="cacd21c8-e424-492a-ba96-b618800b3d58" containerID="bd6b1c78bc526ac62c4474e8e799531a33f0e623e8c0a164bac97d9751a89e1e" exitCode=0 Nov 24 14:32:17 crc kubenswrapper[4822]: I1124 14:32:17.034732 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" event={"ID":"cacd21c8-e424-492a-ba96-b618800b3d58","Type":"ContainerDied","Data":"bd6b1c78bc526ac62c4474e8e799531a33f0e623e8c0a164bac97d9751a89e1e"} Nov 24 14:32:17 crc kubenswrapper[4822]: I1124 14:32:17.037421 4822 generic.go:334] "Generic (PLEG): container finished" podID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerID="2c2494870d03a5e28786cbdfd5aff15a304523b75a5b2b01da2e0619a32e00a4" exitCode=0 Nov 24 14:32:17 crc kubenswrapper[4822]: I1124 14:32:17.037501 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hldt" event={"ID":"8169b355-d354-4eb8-8abc-fe1cc44b8557","Type":"ContainerDied","Data":"2c2494870d03a5e28786cbdfd5aff15a304523b75a5b2b01da2e0619a32e00a4"} Nov 24 14:32:17 crc kubenswrapper[4822]: I1124 14:32:17.037545 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hldt" event={"ID":"8169b355-d354-4eb8-8abc-fe1cc44b8557","Type":"ContainerStarted","Data":"06a365d0eaa53084431bf9dc881c93ea43b2d46d80cc32304cb1b24b502ac70f"} Nov 24 14:32:18 crc kubenswrapper[4822]: I1124 14:32:18.046584 4822 generic.go:334] "Generic (PLEG): container finished" podID="cacd21c8-e424-492a-ba96-b618800b3d58" containerID="ac6ebd45d892a2f54e350fcc5949018c658f8f33c0adbd966f5942ebcb505798" exitCode=0 Nov 24 14:32:18 crc kubenswrapper[4822]: I1124 14:32:18.046643 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" event={"ID":"cacd21c8-e424-492a-ba96-b618800b3d58","Type":"ContainerDied","Data":"ac6ebd45d892a2f54e350fcc5949018c658f8f33c0adbd966f5942ebcb505798"} Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.056288 4822 generic.go:334] "Generic (PLEG): container finished" podID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerID="d9a4b2001c2fe16b4387800ec22bc0111137d33b9f1bdda28abfb1ae5c2f0d31" exitCode=0 Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.056476 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hldt" event={"ID":"8169b355-d354-4eb8-8abc-fe1cc44b8557","Type":"ContainerDied","Data":"d9a4b2001c2fe16b4387800ec22bc0111137d33b9f1bdda28abfb1ae5c2f0d31"} Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.530156 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.683469 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-util\") pod \"cacd21c8-e424-492a-ba96-b618800b3d58\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.683524 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqsf2\" (UniqueName: \"kubernetes.io/projected/cacd21c8-e424-492a-ba96-b618800b3d58-kube-api-access-vqsf2\") pod \"cacd21c8-e424-492a-ba96-b618800b3d58\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.683693 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-bundle\") pod \"cacd21c8-e424-492a-ba96-b618800b3d58\" (UID: \"cacd21c8-e424-492a-ba96-b618800b3d58\") " Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.685009 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-bundle" (OuterVolumeSpecName: "bundle") pod "cacd21c8-e424-492a-ba96-b618800b3d58" (UID: "cacd21c8-e424-492a-ba96-b618800b3d58"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.691123 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cacd21c8-e424-492a-ba96-b618800b3d58-kube-api-access-vqsf2" (OuterVolumeSpecName: "kube-api-access-vqsf2") pod "cacd21c8-e424-492a-ba96-b618800b3d58" (UID: "cacd21c8-e424-492a-ba96-b618800b3d58"). InnerVolumeSpecName "kube-api-access-vqsf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.785449 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqsf2\" (UniqueName: \"kubernetes.io/projected/cacd21c8-e424-492a-ba96-b618800b3d58-kube-api-access-vqsf2\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.785493 4822 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.849540 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-util" (OuterVolumeSpecName: "util") pod "cacd21c8-e424-492a-ba96-b618800b3d58" (UID: "cacd21c8-e424-492a-ba96-b618800b3d58"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:19 crc kubenswrapper[4822]: I1124 14:32:19.887598 4822 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cacd21c8-e424-492a-ba96-b618800b3d58-util\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:20 crc kubenswrapper[4822]: I1124 14:32:20.065916 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" event={"ID":"cacd21c8-e424-492a-ba96-b618800b3d58","Type":"ContainerDied","Data":"6e0fb67d7225c3911beaefcbcbd7304852e50f6f532d9949fcd20fb581d301bb"} Nov 24 14:32:20 crc kubenswrapper[4822]: I1124 14:32:20.065966 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e0fb67d7225c3911beaefcbcbd7304852e50f6f532d9949fcd20fb581d301bb" Nov 24 14:32:20 crc kubenswrapper[4822]: I1124 14:32:20.065981 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94" Nov 24 14:32:21 crc kubenswrapper[4822]: I1124 14:32:21.077185 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hldt" event={"ID":"8169b355-d354-4eb8-8abc-fe1cc44b8557","Type":"ContainerStarted","Data":"45b1d2a26fc2a5fb4adaa467ea49ca69ab038d31e956e2d3c4a4d8811887af42"} Nov 24 14:32:21 crc kubenswrapper[4822]: I1124 14:32:21.104115 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6hldt" podStartSLOduration=2.855686863 podStartE2EDuration="6.104086709s" podCreationTimestamp="2025-11-24 14:32:15 +0000 UTC" firstStartedPulling="2025-11-24 14:32:17.038839942 +0000 UTC m=+774.155480419" lastFinishedPulling="2025-11-24 14:32:20.287239788 +0000 UTC m=+777.403880265" observedRunningTime="2025-11-24 14:32:21.10126266 +0000 UTC m=+778.217903137" watchObservedRunningTime="2025-11-24 14:32:21.104086709 +0000 UTC m=+778.220727186" Nov 24 14:32:25 crc kubenswrapper[4822]: I1124 14:32:25.602706 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:25 crc kubenswrapper[4822]: I1124 14:32:25.603600 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:26 crc kubenswrapper[4822]: I1124 14:32:26.652561 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6hldt" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerName="registry-server" probeResult="failure" output=< Nov 24 14:32:26 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:32:26 crc kubenswrapper[4822]: > Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.766946 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw"] Nov 24 14:32:28 crc kubenswrapper[4822]: E1124 14:32:28.767763 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cacd21c8-e424-492a-ba96-b618800b3d58" containerName="util" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.767778 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="cacd21c8-e424-492a-ba96-b618800b3d58" containerName="util" Nov 24 14:32:28 crc kubenswrapper[4822]: E1124 14:32:28.767810 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cacd21c8-e424-492a-ba96-b618800b3d58" containerName="pull" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.767815 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="cacd21c8-e424-492a-ba96-b618800b3d58" containerName="pull" Nov 24 14:32:28 crc kubenswrapper[4822]: E1124 14:32:28.767830 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cacd21c8-e424-492a-ba96-b618800b3d58" containerName="extract" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.767838 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="cacd21c8-e424-492a-ba96-b618800b3d58" containerName="extract" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.767969 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="cacd21c8-e424-492a-ba96-b618800b3d58" containerName="extract" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.769244 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.772749 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.773008 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-5dkkq" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.773511 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.774779 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.782146 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw"] Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.782371 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.821392 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pnc8\" (UniqueName: \"kubernetes.io/projected/266b3b36-bac6-482f-b35e-5153d0c3b099-kube-api-access-2pnc8\") pod \"metallb-operator-controller-manager-85fcb85fb9-hk6dw\" (UID: \"266b3b36-bac6-482f-b35e-5153d0c3b099\") " pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.821469 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/266b3b36-bac6-482f-b35e-5153d0c3b099-webhook-cert\") pod \"metallb-operator-controller-manager-85fcb85fb9-hk6dw\" (UID: \"266b3b36-bac6-482f-b35e-5153d0c3b099\") " pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.821526 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/266b3b36-bac6-482f-b35e-5153d0c3b099-apiservice-cert\") pod \"metallb-operator-controller-manager-85fcb85fb9-hk6dw\" (UID: \"266b3b36-bac6-482f-b35e-5153d0c3b099\") " pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.922821 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pnc8\" (UniqueName: \"kubernetes.io/projected/266b3b36-bac6-482f-b35e-5153d0c3b099-kube-api-access-2pnc8\") pod \"metallb-operator-controller-manager-85fcb85fb9-hk6dw\" (UID: \"266b3b36-bac6-482f-b35e-5153d0c3b099\") " pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.922881 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/266b3b36-bac6-482f-b35e-5153d0c3b099-webhook-cert\") pod \"metallb-operator-controller-manager-85fcb85fb9-hk6dw\" (UID: \"266b3b36-bac6-482f-b35e-5153d0c3b099\") " pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.922926 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/266b3b36-bac6-482f-b35e-5153d0c3b099-apiservice-cert\") pod \"metallb-operator-controller-manager-85fcb85fb9-hk6dw\" (UID: \"266b3b36-bac6-482f-b35e-5153d0c3b099\") " pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.932314 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/266b3b36-bac6-482f-b35e-5153d0c3b099-webhook-cert\") pod \"metallb-operator-controller-manager-85fcb85fb9-hk6dw\" (UID: \"266b3b36-bac6-482f-b35e-5153d0c3b099\") " pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.932376 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/266b3b36-bac6-482f-b35e-5153d0c3b099-apiservice-cert\") pod \"metallb-operator-controller-manager-85fcb85fb9-hk6dw\" (UID: \"266b3b36-bac6-482f-b35e-5153d0c3b099\") " pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:28 crc kubenswrapper[4822]: I1124 14:32:28.949120 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pnc8\" (UniqueName: \"kubernetes.io/projected/266b3b36-bac6-482f-b35e-5153d0c3b099-kube-api-access-2pnc8\") pod \"metallb-operator-controller-manager-85fcb85fb9-hk6dw\" (UID: \"266b3b36-bac6-482f-b35e-5153d0c3b099\") " pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.099496 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.212138 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp"] Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.213558 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.219381 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.219428 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-99ffk" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.219709 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.249945 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp"] Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.331160 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b14b176-790c-4e60-9f93-bda90fda768a-webhook-cert\") pod \"metallb-operator-webhook-server-5fbf9ff7c-xsgkp\" (UID: \"0b14b176-790c-4e60-9f93-bda90fda768a\") " pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.331237 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b14b176-790c-4e60-9f93-bda90fda768a-apiservice-cert\") pod \"metallb-operator-webhook-server-5fbf9ff7c-xsgkp\" (UID: \"0b14b176-790c-4e60-9f93-bda90fda768a\") " pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.331275 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7z5b\" (UniqueName: \"kubernetes.io/projected/0b14b176-790c-4e60-9f93-bda90fda768a-kube-api-access-q7z5b\") pod \"metallb-operator-webhook-server-5fbf9ff7c-xsgkp\" (UID: \"0b14b176-790c-4e60-9f93-bda90fda768a\") " pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.432789 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b14b176-790c-4e60-9f93-bda90fda768a-webhook-cert\") pod \"metallb-operator-webhook-server-5fbf9ff7c-xsgkp\" (UID: \"0b14b176-790c-4e60-9f93-bda90fda768a\") " pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.432872 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b14b176-790c-4e60-9f93-bda90fda768a-apiservice-cert\") pod \"metallb-operator-webhook-server-5fbf9ff7c-xsgkp\" (UID: \"0b14b176-790c-4e60-9f93-bda90fda768a\") " pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.432904 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7z5b\" (UniqueName: \"kubernetes.io/projected/0b14b176-790c-4e60-9f93-bda90fda768a-kube-api-access-q7z5b\") pod \"metallb-operator-webhook-server-5fbf9ff7c-xsgkp\" (UID: \"0b14b176-790c-4e60-9f93-bda90fda768a\") " pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.449589 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b14b176-790c-4e60-9f93-bda90fda768a-webhook-cert\") pod \"metallb-operator-webhook-server-5fbf9ff7c-xsgkp\" (UID: \"0b14b176-790c-4e60-9f93-bda90fda768a\") " pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.456719 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b14b176-790c-4e60-9f93-bda90fda768a-apiservice-cert\") pod \"metallb-operator-webhook-server-5fbf9ff7c-xsgkp\" (UID: \"0b14b176-790c-4e60-9f93-bda90fda768a\") " pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.463647 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7z5b\" (UniqueName: \"kubernetes.io/projected/0b14b176-790c-4e60-9f93-bda90fda768a-kube-api-access-q7z5b\") pod \"metallb-operator-webhook-server-5fbf9ff7c-xsgkp\" (UID: \"0b14b176-790c-4e60-9f93-bda90fda768a\") " pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.530566 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.735939 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw"] Nov 24 14:32:29 crc kubenswrapper[4822]: I1124 14:32:29.973741 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp"] Nov 24 14:32:29 crc kubenswrapper[4822]: W1124 14:32:29.976498 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b14b176_790c_4e60_9f93_bda90fda768a.slice/crio-3cc21a05c75b238af47e9673c1f3152901a1f1f64617293e7d269c9b0fe9e8bf WatchSource:0}: Error finding container 3cc21a05c75b238af47e9673c1f3152901a1f1f64617293e7d269c9b0fe9e8bf: Status 404 returned error can't find the container with id 3cc21a05c75b238af47e9673c1f3152901a1f1f64617293e7d269c9b0fe9e8bf Nov 24 14:32:30 crc kubenswrapper[4822]: I1124 14:32:30.148040 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" event={"ID":"266b3b36-bac6-482f-b35e-5153d0c3b099","Type":"ContainerStarted","Data":"399ef464387bc6c9f6e34cf15887eb398378eac3fde0cb094f20723bcda3a5c4"} Nov 24 14:32:30 crc kubenswrapper[4822]: I1124 14:32:30.149583 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" event={"ID":"0b14b176-790c-4e60-9f93-bda90fda768a","Type":"ContainerStarted","Data":"3cc21a05c75b238af47e9673c1f3152901a1f1f64617293e7d269c9b0fe9e8bf"} Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.448249 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kl5l2"] Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.450405 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.462165 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kl5l2"] Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.515376 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp2sn\" (UniqueName: \"kubernetes.io/projected/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-kube-api-access-hp2sn\") pod \"community-operators-kl5l2\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.515455 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-utilities\") pod \"community-operators-kl5l2\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.515506 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-catalog-content\") pod \"community-operators-kl5l2\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.642369 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-catalog-content\") pod \"community-operators-kl5l2\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.642768 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp2sn\" (UniqueName: \"kubernetes.io/projected/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-kube-api-access-hp2sn\") pod \"community-operators-kl5l2\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.642999 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-utilities\") pod \"community-operators-kl5l2\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.643690 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-catalog-content\") pod \"community-operators-kl5l2\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.643742 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-utilities\") pod \"community-operators-kl5l2\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.676817 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp2sn\" (UniqueName: \"kubernetes.io/projected/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-kube-api-access-hp2sn\") pod \"community-operators-kl5l2\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:34 crc kubenswrapper[4822]: I1124 14:32:34.778981 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:35 crc kubenswrapper[4822]: I1124 14:32:35.654479 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:35 crc kubenswrapper[4822]: I1124 14:32:35.715743 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.640791 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g4tpw"] Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.642362 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.664101 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g4tpw"] Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.698258 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-utilities\") pod \"redhat-marketplace-g4tpw\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.698335 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jjx2\" (UniqueName: \"kubernetes.io/projected/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-kube-api-access-5jjx2\") pod \"redhat-marketplace-g4tpw\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.698421 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-catalog-content\") pod \"redhat-marketplace-g4tpw\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.760518 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kl5l2"] Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.800792 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jjx2\" (UniqueName: \"kubernetes.io/projected/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-kube-api-access-5jjx2\") pod \"redhat-marketplace-g4tpw\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.800971 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-catalog-content\") pod \"redhat-marketplace-g4tpw\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.801046 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-utilities\") pod \"redhat-marketplace-g4tpw\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.801711 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-utilities\") pod \"redhat-marketplace-g4tpw\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.801713 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-catalog-content\") pod \"redhat-marketplace-g4tpw\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.827330 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jjx2\" (UniqueName: \"kubernetes.io/projected/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-kube-api-access-5jjx2\") pod \"redhat-marketplace-g4tpw\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:37 crc kubenswrapper[4822]: I1124 14:32:37.959313 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.215565 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" event={"ID":"266b3b36-bac6-482f-b35e-5153d0c3b099","Type":"ContainerStarted","Data":"b7812dd4116740490affdd13fada722b937db36ff7411d5d9f0f6c80dedec8b6"} Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.216337 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.222050 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" event={"ID":"0b14b176-790c-4e60-9f93-bda90fda768a","Type":"ContainerStarted","Data":"11c467d11ddeecea22bad78c2530ac0d1fa7f54ee3da7ea1e7271ac746f7b8b4"} Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.222222 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.224135 4822 generic.go:334] "Generic (PLEG): container finished" podID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerID="c1a2bb649764b273fbb73f291529c7b1c09ddf4069b1ac0d2c1ee92c1c42d2d6" exitCode=0 Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.224163 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl5l2" event={"ID":"0557b94c-a5c5-4fbb-9921-c7c355dba4b8","Type":"ContainerDied","Data":"c1a2bb649764b273fbb73f291529c7b1c09ddf4069b1ac0d2c1ee92c1c42d2d6"} Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.224180 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl5l2" event={"ID":"0557b94c-a5c5-4fbb-9921-c7c355dba4b8","Type":"ContainerStarted","Data":"627a77b280e18f0f8dcddf7b78eb92cee28a9ec685b16d0619ab98ccb2f071db"} Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.238501 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" podStartSLOduration=2.749196388 podStartE2EDuration="10.238481723s" podCreationTimestamp="2025-11-24 14:32:28 +0000 UTC" firstStartedPulling="2025-11-24 14:32:29.778051852 +0000 UTC m=+786.894692329" lastFinishedPulling="2025-11-24 14:32:37.267337187 +0000 UTC m=+794.383977664" observedRunningTime="2025-11-24 14:32:38.2365194 +0000 UTC m=+795.353159877" watchObservedRunningTime="2025-11-24 14:32:38.238481723 +0000 UTC m=+795.355122200" Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.270226 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" podStartSLOduration=1.959393114 podStartE2EDuration="9.270187463s" podCreationTimestamp="2025-11-24 14:32:29 +0000 UTC" firstStartedPulling="2025-11-24 14:32:29.979337948 +0000 UTC m=+787.095978425" lastFinishedPulling="2025-11-24 14:32:37.290132297 +0000 UTC m=+794.406772774" observedRunningTime="2025-11-24 14:32:38.269357998 +0000 UTC m=+795.385998475" watchObservedRunningTime="2025-11-24 14:32:38.270187463 +0000 UTC m=+795.386827940" Nov 24 14:32:38 crc kubenswrapper[4822]: I1124 14:32:38.463322 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g4tpw"] Nov 24 14:32:39 crc kubenswrapper[4822]: I1124 14:32:39.232150 4822 generic.go:334] "Generic (PLEG): container finished" podID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerID="116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f" exitCode=0 Nov 24 14:32:39 crc kubenswrapper[4822]: I1124 14:32:39.234355 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g4tpw" event={"ID":"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a","Type":"ContainerDied","Data":"116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f"} Nov 24 14:32:39 crc kubenswrapper[4822]: I1124 14:32:39.234731 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g4tpw" event={"ID":"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a","Type":"ContainerStarted","Data":"bb5eadf950c71f1563c52d430844db61bea6bdecd58b4741f944b8b4e6b9fd23"} Nov 24 14:32:40 crc kubenswrapper[4822]: I1124 14:32:40.247801 4822 generic.go:334] "Generic (PLEG): container finished" podID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerID="91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229" exitCode=0 Nov 24 14:32:40 crc kubenswrapper[4822]: I1124 14:32:40.247846 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g4tpw" event={"ID":"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a","Type":"ContainerDied","Data":"91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229"} Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.037750 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6hldt"] Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.038516 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6hldt" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerName="registry-server" containerID="cri-o://45b1d2a26fc2a5fb4adaa467ea49ca69ab038d31e956e2d3c4a4d8811887af42" gracePeriod=2 Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.257187 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g4tpw" event={"ID":"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a","Type":"ContainerStarted","Data":"6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38"} Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.260526 4822 generic.go:334] "Generic (PLEG): container finished" podID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerID="45b1d2a26fc2a5fb4adaa467ea49ca69ab038d31e956e2d3c4a4d8811887af42" exitCode=0 Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.260571 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hldt" event={"ID":"8169b355-d354-4eb8-8abc-fe1cc44b8557","Type":"ContainerDied","Data":"45b1d2a26fc2a5fb4adaa467ea49ca69ab038d31e956e2d3c4a4d8811887af42"} Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.287444 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g4tpw" podStartSLOduration=2.847630543 podStartE2EDuration="4.287412687s" podCreationTimestamp="2025-11-24 14:32:37 +0000 UTC" firstStartedPulling="2025-11-24 14:32:39.235931978 +0000 UTC m=+796.352572455" lastFinishedPulling="2025-11-24 14:32:40.675714112 +0000 UTC m=+797.792354599" observedRunningTime="2025-11-24 14:32:41.281023365 +0000 UTC m=+798.397663892" watchObservedRunningTime="2025-11-24 14:32:41.287412687 +0000 UTC m=+798.404053204" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.451962 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-49nkq"] Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.453566 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.478201 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49nkq"] Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.557780 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-catalog-content\") pod \"certified-operators-49nkq\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.557918 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-utilities\") pod \"certified-operators-49nkq\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.557970 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpw5n\" (UniqueName: \"kubernetes.io/projected/9dfcefe0-790f-411b-9c43-5a1665a98cc2-kube-api-access-gpw5n\") pod \"certified-operators-49nkq\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.659429 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-catalog-content\") pod \"certified-operators-49nkq\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.659548 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-utilities\") pod \"certified-operators-49nkq\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.659580 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpw5n\" (UniqueName: \"kubernetes.io/projected/9dfcefe0-790f-411b-9c43-5a1665a98cc2-kube-api-access-gpw5n\") pod \"certified-operators-49nkq\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.660093 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-catalog-content\") pod \"certified-operators-49nkq\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.660365 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-utilities\") pod \"certified-operators-49nkq\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.682589 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpw5n\" (UniqueName: \"kubernetes.io/projected/9dfcefe0-790f-411b-9c43-5a1665a98cc2-kube-api-access-gpw5n\") pod \"certified-operators-49nkq\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:41 crc kubenswrapper[4822]: I1124 14:32:41.775365 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.054565 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.158849 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49nkq"] Nov 24 14:32:42 crc kubenswrapper[4822]: W1124 14:32:42.171674 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dfcefe0_790f_411b_9c43_5a1665a98cc2.slice/crio-b82bc528f0d66ff162d5ad568334a8e878793d86c5270e87ccd1d6ce81ece3c4 WatchSource:0}: Error finding container b82bc528f0d66ff162d5ad568334a8e878793d86c5270e87ccd1d6ce81ece3c4: Status 404 returned error can't find the container with id b82bc528f0d66ff162d5ad568334a8e878793d86c5270e87ccd1d6ce81ece3c4 Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.173655 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs6cg\" (UniqueName: \"kubernetes.io/projected/8169b355-d354-4eb8-8abc-fe1cc44b8557-kube-api-access-rs6cg\") pod \"8169b355-d354-4eb8-8abc-fe1cc44b8557\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.173740 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-utilities\") pod \"8169b355-d354-4eb8-8abc-fe1cc44b8557\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.173818 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-catalog-content\") pod \"8169b355-d354-4eb8-8abc-fe1cc44b8557\" (UID: \"8169b355-d354-4eb8-8abc-fe1cc44b8557\") " Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.175187 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-utilities" (OuterVolumeSpecName: "utilities") pod "8169b355-d354-4eb8-8abc-fe1cc44b8557" (UID: "8169b355-d354-4eb8-8abc-fe1cc44b8557"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.179836 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8169b355-d354-4eb8-8abc-fe1cc44b8557-kube-api-access-rs6cg" (OuterVolumeSpecName: "kube-api-access-rs6cg") pod "8169b355-d354-4eb8-8abc-fe1cc44b8557" (UID: "8169b355-d354-4eb8-8abc-fe1cc44b8557"). InnerVolumeSpecName "kube-api-access-rs6cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.284847 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49nkq" event={"ID":"9dfcefe0-790f-411b-9c43-5a1665a98cc2","Type":"ContainerStarted","Data":"b82bc528f0d66ff162d5ad568334a8e878793d86c5270e87ccd1d6ce81ece3c4"} Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.292529 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.292604 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs6cg\" (UniqueName: \"kubernetes.io/projected/8169b355-d354-4eb8-8abc-fe1cc44b8557-kube-api-access-rs6cg\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.292716 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hldt" event={"ID":"8169b355-d354-4eb8-8abc-fe1cc44b8557","Type":"ContainerDied","Data":"06a365d0eaa53084431bf9dc881c93ea43b2d46d80cc32304cb1b24b502ac70f"} Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.292764 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hldt" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.292807 4822 scope.go:117] "RemoveContainer" containerID="45b1d2a26fc2a5fb4adaa467ea49ca69ab038d31e956e2d3c4a4d8811887af42" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.318514 4822 scope.go:117] "RemoveContainer" containerID="d9a4b2001c2fe16b4387800ec22bc0111137d33b9f1bdda28abfb1ae5c2f0d31" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.345224 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8169b355-d354-4eb8-8abc-fe1cc44b8557" (UID: "8169b355-d354-4eb8-8abc-fe1cc44b8557"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.353930 4822 scope.go:117] "RemoveContainer" containerID="2c2494870d03a5e28786cbdfd5aff15a304523b75a5b2b01da2e0619a32e00a4" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.395722 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8169b355-d354-4eb8-8abc-fe1cc44b8557-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.629504 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6hldt"] Nov 24 14:32:42 crc kubenswrapper[4822]: I1124 14:32:42.633816 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6hldt"] Nov 24 14:32:43 crc kubenswrapper[4822]: I1124 14:32:43.300938 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl5l2" event={"ID":"0557b94c-a5c5-4fbb-9921-c7c355dba4b8","Type":"ContainerStarted","Data":"96a49a3e821d7ceac6c01f0bf2a5359ef72ca8bdfd70aad08bbcae1cb5296f9b"} Nov 24 14:32:43 crc kubenswrapper[4822]: I1124 14:32:43.302343 4822 generic.go:334] "Generic (PLEG): container finished" podID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerID="7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61" exitCode=0 Nov 24 14:32:43 crc kubenswrapper[4822]: I1124 14:32:43.302429 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49nkq" event={"ID":"9dfcefe0-790f-411b-9c43-5a1665a98cc2","Type":"ContainerDied","Data":"7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61"} Nov 24 14:32:43 crc kubenswrapper[4822]: I1124 14:32:43.715121 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" path="/var/lib/kubelet/pods/8169b355-d354-4eb8-8abc-fe1cc44b8557/volumes" Nov 24 14:32:44 crc kubenswrapper[4822]: I1124 14:32:44.317396 4822 generic.go:334] "Generic (PLEG): container finished" podID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerID="96a49a3e821d7ceac6c01f0bf2a5359ef72ca8bdfd70aad08bbcae1cb5296f9b" exitCode=0 Nov 24 14:32:44 crc kubenswrapper[4822]: I1124 14:32:44.317546 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl5l2" event={"ID":"0557b94c-a5c5-4fbb-9921-c7c355dba4b8","Type":"ContainerDied","Data":"96a49a3e821d7ceac6c01f0bf2a5359ef72ca8bdfd70aad08bbcae1cb5296f9b"} Nov 24 14:32:44 crc kubenswrapper[4822]: I1124 14:32:44.325707 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49nkq" event={"ID":"9dfcefe0-790f-411b-9c43-5a1665a98cc2","Type":"ContainerStarted","Data":"fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda"} Nov 24 14:32:45 crc kubenswrapper[4822]: I1124 14:32:45.337010 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl5l2" event={"ID":"0557b94c-a5c5-4fbb-9921-c7c355dba4b8","Type":"ContainerStarted","Data":"6e5518eb635f5a514fff0e52c9b876caaad99d260c05bace5b4cdf88f764c7e9"} Nov 24 14:32:45 crc kubenswrapper[4822]: I1124 14:32:45.340142 4822 generic.go:334] "Generic (PLEG): container finished" podID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerID="fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda" exitCode=0 Nov 24 14:32:45 crc kubenswrapper[4822]: I1124 14:32:45.340240 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49nkq" event={"ID":"9dfcefe0-790f-411b-9c43-5a1665a98cc2","Type":"ContainerDied","Data":"fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda"} Nov 24 14:32:45 crc kubenswrapper[4822]: I1124 14:32:45.398011 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kl5l2" podStartSLOduration=4.884389726 podStartE2EDuration="11.397989894s" podCreationTimestamp="2025-11-24 14:32:34 +0000 UTC" firstStartedPulling="2025-11-24 14:32:38.227756303 +0000 UTC m=+795.344396780" lastFinishedPulling="2025-11-24 14:32:44.741356471 +0000 UTC m=+801.857996948" observedRunningTime="2025-11-24 14:32:45.392955266 +0000 UTC m=+802.509595753" watchObservedRunningTime="2025-11-24 14:32:45.397989894 +0000 UTC m=+802.514630371" Nov 24 14:32:46 crc kubenswrapper[4822]: I1124 14:32:46.349934 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49nkq" event={"ID":"9dfcefe0-790f-411b-9c43-5a1665a98cc2","Type":"ContainerStarted","Data":"2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47"} Nov 24 14:32:47 crc kubenswrapper[4822]: I1124 14:32:47.960067 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:47 crc kubenswrapper[4822]: I1124 14:32:47.960188 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:48 crc kubenswrapper[4822]: I1124 14:32:48.014807 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:48 crc kubenswrapper[4822]: I1124 14:32:48.039043 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-49nkq" podStartSLOduration=4.530497868 podStartE2EDuration="7.039017409s" podCreationTimestamp="2025-11-24 14:32:41 +0000 UTC" firstStartedPulling="2025-11-24 14:32:43.303841748 +0000 UTC m=+800.420482215" lastFinishedPulling="2025-11-24 14:32:45.812361279 +0000 UTC m=+802.929001756" observedRunningTime="2025-11-24 14:32:46.388443309 +0000 UTC m=+803.505083796" watchObservedRunningTime="2025-11-24 14:32:48.039017409 +0000 UTC m=+805.155657906" Nov 24 14:32:48 crc kubenswrapper[4822]: I1124 14:32:48.428747 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:49 crc kubenswrapper[4822]: I1124 14:32:49.535802 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5fbf9ff7c-xsgkp" Nov 24 14:32:51 crc kubenswrapper[4822]: I1124 14:32:51.437004 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g4tpw"] Nov 24 14:32:51 crc kubenswrapper[4822]: I1124 14:32:51.438644 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g4tpw" podUID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerName="registry-server" containerID="cri-o://6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38" gracePeriod=2 Nov 24 14:32:51 crc kubenswrapper[4822]: I1124 14:32:51.776676 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:51 crc kubenswrapper[4822]: I1124 14:32:51.777744 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:51 crc kubenswrapper[4822]: I1124 14:32:51.837156 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:51 crc kubenswrapper[4822]: I1124 14:32:51.890533 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.034532 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-catalog-content\") pod \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.034982 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-utilities\") pod \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.035109 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jjx2\" (UniqueName: \"kubernetes.io/projected/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-kube-api-access-5jjx2\") pod \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\" (UID: \"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a\") " Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.035833 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-utilities" (OuterVolumeSpecName: "utilities") pod "e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" (UID: "e0ab8a86-a991-47f6-a1b2-38a7dabacb2a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.041464 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-kube-api-access-5jjx2" (OuterVolumeSpecName: "kube-api-access-5jjx2") pod "e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" (UID: "e0ab8a86-a991-47f6-a1b2-38a7dabacb2a"). InnerVolumeSpecName "kube-api-access-5jjx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.060151 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" (UID: "e0ab8a86-a991-47f6-a1b2-38a7dabacb2a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.136986 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.137040 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.137053 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jjx2\" (UniqueName: \"kubernetes.io/projected/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a-kube-api-access-5jjx2\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.392191 4822 generic.go:334] "Generic (PLEG): container finished" podID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerID="6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38" exitCode=0 Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.392279 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g4tpw" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.392262 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g4tpw" event={"ID":"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a","Type":"ContainerDied","Data":"6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38"} Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.392506 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g4tpw" event={"ID":"e0ab8a86-a991-47f6-a1b2-38a7dabacb2a","Type":"ContainerDied","Data":"bb5eadf950c71f1563c52d430844db61bea6bdecd58b4741f944b8b4e6b9fd23"} Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.392533 4822 scope.go:117] "RemoveContainer" containerID="6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.422252 4822 scope.go:117] "RemoveContainer" containerID="91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.439049 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g4tpw"] Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.447532 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g4tpw"] Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.457888 4822 scope.go:117] "RemoveContainer" containerID="116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.458052 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.480823 4822 scope.go:117] "RemoveContainer" containerID="6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38" Nov 24 14:32:52 crc kubenswrapper[4822]: E1124 14:32:52.481336 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38\": container with ID starting with 6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38 not found: ID does not exist" containerID="6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.481369 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38"} err="failed to get container status \"6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38\": rpc error: code = NotFound desc = could not find container \"6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38\": container with ID starting with 6baba944df48b91506bbf82f394425b89b9e17ba1c891ccec63b658ba6568c38 not found: ID does not exist" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.481391 4822 scope.go:117] "RemoveContainer" containerID="91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229" Nov 24 14:32:52 crc kubenswrapper[4822]: E1124 14:32:52.481614 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229\": container with ID starting with 91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229 not found: ID does not exist" containerID="91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.481635 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229"} err="failed to get container status \"91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229\": rpc error: code = NotFound desc = could not find container \"91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229\": container with ID starting with 91ed9ada2c638e4ea413369c254deea952eeee31ce1302f81929810f2daf3229 not found: ID does not exist" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.481646 4822 scope.go:117] "RemoveContainer" containerID="116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f" Nov 24 14:32:52 crc kubenswrapper[4822]: E1124 14:32:52.481980 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f\": container with ID starting with 116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f not found: ID does not exist" containerID="116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f" Nov 24 14:32:52 crc kubenswrapper[4822]: I1124 14:32:52.482000 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f"} err="failed to get container status \"116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f\": rpc error: code = NotFound desc = could not find container \"116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f\": container with ID starting with 116995f8d402b7b90f248ef88f8f0508e62bd2b17f9dcff6ec39be0b7c41396f not found: ID does not exist" Nov 24 14:32:53 crc kubenswrapper[4822]: I1124 14:32:53.712594 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" path="/var/lib/kubelet/pods/e0ab8a86-a991-47f6-a1b2-38a7dabacb2a/volumes" Nov 24 14:32:54 crc kubenswrapper[4822]: I1124 14:32:54.780068 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:54 crc kubenswrapper[4822]: I1124 14:32:54.780462 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:54 crc kubenswrapper[4822]: I1124 14:32:54.839338 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:55 crc kubenswrapper[4822]: I1124 14:32:55.432863 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-49nkq"] Nov 24 14:32:55 crc kubenswrapper[4822]: I1124 14:32:55.433159 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-49nkq" podUID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerName="registry-server" containerID="cri-o://2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47" gracePeriod=2 Nov 24 14:32:55 crc kubenswrapper[4822]: I1124 14:32:55.461070 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:55 crc kubenswrapper[4822]: I1124 14:32:55.861626 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:55 crc kubenswrapper[4822]: I1124 14:32:55.993221 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-catalog-content\") pod \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " Nov 24 14:32:55 crc kubenswrapper[4822]: I1124 14:32:55.993421 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-utilities\") pod \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " Nov 24 14:32:55 crc kubenswrapper[4822]: I1124 14:32:55.993462 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpw5n\" (UniqueName: \"kubernetes.io/projected/9dfcefe0-790f-411b-9c43-5a1665a98cc2-kube-api-access-gpw5n\") pod \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\" (UID: \"9dfcefe0-790f-411b-9c43-5a1665a98cc2\") " Nov 24 14:32:55 crc kubenswrapper[4822]: I1124 14:32:55.996278 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-utilities" (OuterVolumeSpecName: "utilities") pod "9dfcefe0-790f-411b-9c43-5a1665a98cc2" (UID: "9dfcefe0-790f-411b-9c43-5a1665a98cc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.008596 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dfcefe0-790f-411b-9c43-5a1665a98cc2-kube-api-access-gpw5n" (OuterVolumeSpecName: "kube-api-access-gpw5n") pod "9dfcefe0-790f-411b-9c43-5a1665a98cc2" (UID: "9dfcefe0-790f-411b-9c43-5a1665a98cc2"). InnerVolumeSpecName "kube-api-access-gpw5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.049446 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9dfcefe0-790f-411b-9c43-5a1665a98cc2" (UID: "9dfcefe0-790f-411b-9c43-5a1665a98cc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.095570 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.095612 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dfcefe0-790f-411b-9c43-5a1665a98cc2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.095626 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpw5n\" (UniqueName: \"kubernetes.io/projected/9dfcefe0-790f-411b-9c43-5a1665a98cc2-kube-api-access-gpw5n\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.424937 4822 generic.go:334] "Generic (PLEG): container finished" podID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerID="2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47" exitCode=0 Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.425009 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49nkq" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.425034 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49nkq" event={"ID":"9dfcefe0-790f-411b-9c43-5a1665a98cc2","Type":"ContainerDied","Data":"2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47"} Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.425340 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49nkq" event={"ID":"9dfcefe0-790f-411b-9c43-5a1665a98cc2","Type":"ContainerDied","Data":"b82bc528f0d66ff162d5ad568334a8e878793d86c5270e87ccd1d6ce81ece3c4"} Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.425380 4822 scope.go:117] "RemoveContainer" containerID="2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.447887 4822 scope.go:117] "RemoveContainer" containerID="fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.461010 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-49nkq"] Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.467944 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-49nkq"] Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.489919 4822 scope.go:117] "RemoveContainer" containerID="7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.510354 4822 scope.go:117] "RemoveContainer" containerID="2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47" Nov 24 14:32:56 crc kubenswrapper[4822]: E1124 14:32:56.511433 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47\": container with ID starting with 2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47 not found: ID does not exist" containerID="2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.511473 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47"} err="failed to get container status \"2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47\": rpc error: code = NotFound desc = could not find container \"2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47\": container with ID starting with 2f7cc75d2fe6f7740e9519921c220b67b7f4b61425f17bfda9537477ce68bd47 not found: ID does not exist" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.511502 4822 scope.go:117] "RemoveContainer" containerID="fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda" Nov 24 14:32:56 crc kubenswrapper[4822]: E1124 14:32:56.512008 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda\": container with ID starting with fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda not found: ID does not exist" containerID="fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.512076 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda"} err="failed to get container status \"fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda\": rpc error: code = NotFound desc = could not find container \"fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda\": container with ID starting with fd72df4eb832c5e35482c34b739921ea325383f5f86f82e284ce0ad145e5fcda not found: ID does not exist" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.512134 4822 scope.go:117] "RemoveContainer" containerID="7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61" Nov 24 14:32:56 crc kubenswrapper[4822]: E1124 14:32:56.512893 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61\": container with ID starting with 7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61 not found: ID does not exist" containerID="7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61" Nov 24 14:32:56 crc kubenswrapper[4822]: I1124 14:32:56.512925 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61"} err="failed to get container status \"7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61\": rpc error: code = NotFound desc = could not find container \"7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61\": container with ID starting with 7683f3371969564ac295047aec74b5292419778f524744b9584fef5187b61d61 not found: ID does not exist" Nov 24 14:32:57 crc kubenswrapper[4822]: I1124 14:32:57.715613 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" path="/var/lib/kubelet/pods/9dfcefe0-790f-411b-9c43-5a1665a98cc2/volumes" Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.231488 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kl5l2"] Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.231806 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kl5l2" podUID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerName="registry-server" containerID="cri-o://6e5518eb635f5a514fff0e52c9b876caaad99d260c05bace5b4cdf88f764c7e9" gracePeriod=2 Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.476665 4822 generic.go:334] "Generic (PLEG): container finished" podID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerID="6e5518eb635f5a514fff0e52c9b876caaad99d260c05bace5b4cdf88f764c7e9" exitCode=0 Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.477114 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl5l2" event={"ID":"0557b94c-a5c5-4fbb-9921-c7c355dba4b8","Type":"ContainerDied","Data":"6e5518eb635f5a514fff0e52c9b876caaad99d260c05bace5b4cdf88f764c7e9"} Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.606058 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.731860 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp2sn\" (UniqueName: \"kubernetes.io/projected/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-kube-api-access-hp2sn\") pod \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.732251 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-utilities\") pod \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.732421 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-catalog-content\") pod \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\" (UID: \"0557b94c-a5c5-4fbb-9921-c7c355dba4b8\") " Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.733080 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-utilities" (OuterVolumeSpecName: "utilities") pod "0557b94c-a5c5-4fbb-9921-c7c355dba4b8" (UID: "0557b94c-a5c5-4fbb-9921-c7c355dba4b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.741591 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-kube-api-access-hp2sn" (OuterVolumeSpecName: "kube-api-access-hp2sn") pod "0557b94c-a5c5-4fbb-9921-c7c355dba4b8" (UID: "0557b94c-a5c5-4fbb-9921-c7c355dba4b8"). InnerVolumeSpecName "kube-api-access-hp2sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.801834 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0557b94c-a5c5-4fbb-9921-c7c355dba4b8" (UID: "0557b94c-a5c5-4fbb-9921-c7c355dba4b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.834934 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp2sn\" (UniqueName: \"kubernetes.io/projected/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-kube-api-access-hp2sn\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.835198 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:58 crc kubenswrapper[4822]: I1124 14:32:58.835272 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0557b94c-a5c5-4fbb-9921-c7c355dba4b8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:59 crc kubenswrapper[4822]: I1124 14:32:59.486431 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl5l2" event={"ID":"0557b94c-a5c5-4fbb-9921-c7c355dba4b8","Type":"ContainerDied","Data":"627a77b280e18f0f8dcddf7b78eb92cee28a9ec685b16d0619ab98ccb2f071db"} Nov 24 14:32:59 crc kubenswrapper[4822]: I1124 14:32:59.486543 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kl5l2" Nov 24 14:32:59 crc kubenswrapper[4822]: I1124 14:32:59.486972 4822 scope.go:117] "RemoveContainer" containerID="6e5518eb635f5a514fff0e52c9b876caaad99d260c05bace5b4cdf88f764c7e9" Nov 24 14:32:59 crc kubenswrapper[4822]: I1124 14:32:59.507699 4822 scope.go:117] "RemoveContainer" containerID="96a49a3e821d7ceac6c01f0bf2a5359ef72ca8bdfd70aad08bbcae1cb5296f9b" Nov 24 14:32:59 crc kubenswrapper[4822]: I1124 14:32:59.523984 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kl5l2"] Nov 24 14:32:59 crc kubenswrapper[4822]: I1124 14:32:59.533129 4822 scope.go:117] "RemoveContainer" containerID="c1a2bb649764b273fbb73f291529c7b1c09ddf4069b1ac0d2c1ee92c1c42d2d6" Nov 24 14:32:59 crc kubenswrapper[4822]: I1124 14:32:59.533689 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kl5l2"] Nov 24 14:32:59 crc kubenswrapper[4822]: I1124 14:32:59.718029 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" path="/var/lib/kubelet/pods/0557b94c-a5c5-4fbb-9921-c7c355dba4b8/volumes" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.105478 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-85fcb85fb9-hk6dw" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897333 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9ld68"] Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897684 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerName="extract-content" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897705 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerName="extract-content" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897720 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerName="extract-content" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897726 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerName="extract-content" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897738 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerName="extract-content" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897745 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerName="extract-content" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897752 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerName="extract-content" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897758 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerName="extract-content" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897768 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerName="extract-utilities" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897775 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerName="extract-utilities" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897784 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897790 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897799 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897806 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897818 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897824 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897833 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerName="extract-utilities" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897838 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerName="extract-utilities" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897845 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerName="extract-utilities" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897851 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerName="extract-utilities" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897860 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerName="extract-utilities" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897866 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerName="extract-utilities" Nov 24 14:33:09 crc kubenswrapper[4822]: E1124 14:33:09.897875 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897881 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.897998 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ab8a86-a991-47f6-a1b2-38a7dabacb2a" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.898015 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dfcefe0-790f-411b-9c43-5a1665a98cc2" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.898026 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="0557b94c-a5c5-4fbb-9921-c7c355dba4b8" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.898035 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="8169b355-d354-4eb8-8abc-fe1cc44b8557" containerName="registry-server" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.898598 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.901146 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ts9cq"] Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.901593 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.901989 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-zgwd7" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.904429 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.907947 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.908272 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 14:33:09 crc kubenswrapper[4822]: I1124 14:33:09.934904 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9ld68"] Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.023406 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qd7vl"] Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.024804 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.029098 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.029367 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.029446 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kj65l" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.029475 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-dlvwp"] Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.030185 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.030780 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.035041 4822 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.038394 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-reloader\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.038433 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6ca85134-1017-4276-85a1-ac5cf4c47c0b-frr-startup\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.038456 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdqms\" (UniqueName: \"kubernetes.io/projected/7144b776-5528-4a9a-9489-6e73c2608485-kube-api-access-cdqms\") pod \"frr-k8s-webhook-server-6998585d5-9ld68\" (UID: \"7144b776-5528-4a9a-9489-6e73c2608485\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.038482 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-frr-sockets\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.038504 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7144b776-5528-4a9a-9489-6e73c2608485-cert\") pod \"frr-k8s-webhook-server-6998585d5-9ld68\" (UID: \"7144b776-5528-4a9a-9489-6e73c2608485\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.038531 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-frr-conf\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.038556 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbms9\" (UniqueName: \"kubernetes.io/projected/6ca85134-1017-4276-85a1-ac5cf4c47c0b-kube-api-access-tbms9\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.038575 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-metrics\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.038601 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ca85134-1017-4276-85a1-ac5cf4c47c0b-metrics-certs\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.052963 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-dlvwp"] Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.141005 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-frr-sockets\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.141175 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7144b776-5528-4a9a-9489-6e73c2608485-cert\") pod \"frr-k8s-webhook-server-6998585d5-9ld68\" (UID: \"7144b776-5528-4a9a-9489-6e73c2608485\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.141270 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6r9w\" (UniqueName: \"kubernetes.io/projected/2d55bf7d-877f-4605-9eea-b792b465e271-kube-api-access-h6r9w\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.141366 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-frr-conf\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.141440 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed1304c2-b03d-45de-a7c2-dfc08da3691d-cert\") pod \"controller-6c7b4b5f48-dlvwp\" (UID: \"ed1304c2-b03d-45de-a7c2-dfc08da3691d\") " pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.141450 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-frr-sockets\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142098 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-metrics\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142153 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbms9\" (UniqueName: \"kubernetes.io/projected/6ca85134-1017-4276-85a1-ac5cf4c47c0b-kube-api-access-tbms9\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142180 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-memberlist\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142286 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ca85134-1017-4276-85a1-ac5cf4c47c0b-metrics-certs\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142350 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lctqf\" (UniqueName: \"kubernetes.io/projected/ed1304c2-b03d-45de-a7c2-dfc08da3691d-kube-api-access-lctqf\") pod \"controller-6c7b4b5f48-dlvwp\" (UID: \"ed1304c2-b03d-45de-a7c2-dfc08da3691d\") " pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142423 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-frr-conf\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142490 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-reloader\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142519 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2d55bf7d-877f-4605-9eea-b792b465e271-metallb-excludel2\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142566 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ed1304c2-b03d-45de-a7c2-dfc08da3691d-metrics-certs\") pod \"controller-6c7b4b5f48-dlvwp\" (UID: \"ed1304c2-b03d-45de-a7c2-dfc08da3691d\") " pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142614 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6ca85134-1017-4276-85a1-ac5cf4c47c0b-frr-startup\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142651 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-metrics-certs\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142684 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdqms\" (UniqueName: \"kubernetes.io/projected/7144b776-5528-4a9a-9489-6e73c2608485-kube-api-access-cdqms\") pod \"frr-k8s-webhook-server-6998585d5-9ld68\" (UID: \"7144b776-5528-4a9a-9489-6e73c2608485\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.142702 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-metrics\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.143369 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6ca85134-1017-4276-85a1-ac5cf4c47c0b-reloader\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.147058 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6ca85134-1017-4276-85a1-ac5cf4c47c0b-frr-startup\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.151578 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7144b776-5528-4a9a-9489-6e73c2608485-cert\") pod \"frr-k8s-webhook-server-6998585d5-9ld68\" (UID: \"7144b776-5528-4a9a-9489-6e73c2608485\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.164054 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbms9\" (UniqueName: \"kubernetes.io/projected/6ca85134-1017-4276-85a1-ac5cf4c47c0b-kube-api-access-tbms9\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.166496 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdqms\" (UniqueName: \"kubernetes.io/projected/7144b776-5528-4a9a-9489-6e73c2608485-kube-api-access-cdqms\") pod \"frr-k8s-webhook-server-6998585d5-9ld68\" (UID: \"7144b776-5528-4a9a-9489-6e73c2608485\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.167816 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ca85134-1017-4276-85a1-ac5cf4c47c0b-metrics-certs\") pod \"frr-k8s-ts9cq\" (UID: \"6ca85134-1017-4276-85a1-ac5cf4c47c0b\") " pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.244431 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2d55bf7d-877f-4605-9eea-b792b465e271-metallb-excludel2\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.244494 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ed1304c2-b03d-45de-a7c2-dfc08da3691d-metrics-certs\") pod \"controller-6c7b4b5f48-dlvwp\" (UID: \"ed1304c2-b03d-45de-a7c2-dfc08da3691d\") " pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.244531 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-metrics-certs\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.244572 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6r9w\" (UniqueName: \"kubernetes.io/projected/2d55bf7d-877f-4605-9eea-b792b465e271-kube-api-access-h6r9w\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.244602 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed1304c2-b03d-45de-a7c2-dfc08da3691d-cert\") pod \"controller-6c7b4b5f48-dlvwp\" (UID: \"ed1304c2-b03d-45de-a7c2-dfc08da3691d\") " pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.244635 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-memberlist\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.244661 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lctqf\" (UniqueName: \"kubernetes.io/projected/ed1304c2-b03d-45de-a7c2-dfc08da3691d-kube-api-access-lctqf\") pod \"controller-6c7b4b5f48-dlvwp\" (UID: \"ed1304c2-b03d-45de-a7c2-dfc08da3691d\") " pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.245610 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2d55bf7d-877f-4605-9eea-b792b465e271-metallb-excludel2\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: E1124 14:33:10.245715 4822 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 14:33:10 crc kubenswrapper[4822]: E1124 14:33:10.245770 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-memberlist podName:2d55bf7d-877f-4605-9eea-b792b465e271 nodeName:}" failed. No retries permitted until 2025-11-24 14:33:10.7457521 +0000 UTC m=+827.862392577 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-memberlist") pod "speaker-qd7vl" (UID: "2d55bf7d-877f-4605-9eea-b792b465e271") : secret "metallb-memberlist" not found Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.246024 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.248677 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed1304c2-b03d-45de-a7c2-dfc08da3691d-cert\") pod \"controller-6c7b4b5f48-dlvwp\" (UID: \"ed1304c2-b03d-45de-a7c2-dfc08da3691d\") " pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.259024 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ed1304c2-b03d-45de-a7c2-dfc08da3691d-metrics-certs\") pod \"controller-6c7b4b5f48-dlvwp\" (UID: \"ed1304c2-b03d-45de-a7c2-dfc08da3691d\") " pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.264369 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-metrics-certs\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.271105 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6r9w\" (UniqueName: \"kubernetes.io/projected/2d55bf7d-877f-4605-9eea-b792b465e271-kube-api-access-h6r9w\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.271433 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lctqf\" (UniqueName: \"kubernetes.io/projected/ed1304c2-b03d-45de-a7c2-dfc08da3691d-kube-api-access-lctqf\") pod \"controller-6c7b4b5f48-dlvwp\" (UID: \"ed1304c2-b03d-45de-a7c2-dfc08da3691d\") " pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.273319 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.361290 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.570983 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerStarted","Data":"6e6b2166e01fd2d04ebbc28ca16f00e8b5e43585e656fdafab46eef9894bdaea"} Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.728306 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9ld68"] Nov 24 14:33:10 crc kubenswrapper[4822]: W1124 14:33:10.737128 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7144b776_5528_4a9a_9489_6e73c2608485.slice/crio-64e391e95c84a07ac8022f47256791174219439751d3c5cba684d012c24ab6fb WatchSource:0}: Error finding container 64e391e95c84a07ac8022f47256791174219439751d3c5cba684d012c24ab6fb: Status 404 returned error can't find the container with id 64e391e95c84a07ac8022f47256791174219439751d3c5cba684d012c24ab6fb Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.753573 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-memberlist\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:10 crc kubenswrapper[4822]: E1124 14:33:10.753847 4822 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 14:33:10 crc kubenswrapper[4822]: E1124 14:33:10.753978 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-memberlist podName:2d55bf7d-877f-4605-9eea-b792b465e271 nodeName:}" failed. No retries permitted until 2025-11-24 14:33:11.753950096 +0000 UTC m=+828.870590583 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-memberlist") pod "speaker-qd7vl" (UID: "2d55bf7d-877f-4605-9eea-b792b465e271") : secret "metallb-memberlist" not found Nov 24 14:33:10 crc kubenswrapper[4822]: I1124 14:33:10.812577 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-dlvwp"] Nov 24 14:33:10 crc kubenswrapper[4822]: W1124 14:33:10.818380 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded1304c2_b03d_45de_a7c2_dfc08da3691d.slice/crio-38b63b1fcde4a7989326aa9e86da75d690af163b5ba95d3bb38c5eb8ab45f7e7 WatchSource:0}: Error finding container 38b63b1fcde4a7989326aa9e86da75d690af163b5ba95d3bb38c5eb8ab45f7e7: Status 404 returned error can't find the container with id 38b63b1fcde4a7989326aa9e86da75d690af163b5ba95d3bb38c5eb8ab45f7e7 Nov 24 14:33:11 crc kubenswrapper[4822]: I1124 14:33:11.583459 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-dlvwp" event={"ID":"ed1304c2-b03d-45de-a7c2-dfc08da3691d","Type":"ContainerStarted","Data":"714907997e10019d36333586a02131cca1854c65985e357f56a5519af7c11fee"} Nov 24 14:33:11 crc kubenswrapper[4822]: I1124 14:33:11.583836 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-dlvwp" event={"ID":"ed1304c2-b03d-45de-a7c2-dfc08da3691d","Type":"ContainerStarted","Data":"3b80fe9551b85ec3fb11e4c5a7e40236ce2361f0275986d7225a34563c76f1ed"} Nov 24 14:33:11 crc kubenswrapper[4822]: I1124 14:33:11.583861 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-dlvwp" event={"ID":"ed1304c2-b03d-45de-a7c2-dfc08da3691d","Type":"ContainerStarted","Data":"38b63b1fcde4a7989326aa9e86da75d690af163b5ba95d3bb38c5eb8ab45f7e7"} Nov 24 14:33:11 crc kubenswrapper[4822]: I1124 14:33:11.584486 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:11 crc kubenswrapper[4822]: I1124 14:33:11.586077 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" event={"ID":"7144b776-5528-4a9a-9489-6e73c2608485","Type":"ContainerStarted","Data":"64e391e95c84a07ac8022f47256791174219439751d3c5cba684d012c24ab6fb"} Nov 24 14:33:11 crc kubenswrapper[4822]: I1124 14:33:11.607219 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-dlvwp" podStartSLOduration=2.607177227 podStartE2EDuration="2.607177227s" podCreationTimestamp="2025-11-24 14:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:33:11.603224263 +0000 UTC m=+828.719864750" watchObservedRunningTime="2025-11-24 14:33:11.607177227 +0000 UTC m=+828.723817704" Nov 24 14:33:11 crc kubenswrapper[4822]: I1124 14:33:11.769799 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-memberlist\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:11 crc kubenswrapper[4822]: I1124 14:33:11.780888 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2d55bf7d-877f-4605-9eea-b792b465e271-memberlist\") pod \"speaker-qd7vl\" (UID: \"2d55bf7d-877f-4605-9eea-b792b465e271\") " pod="metallb-system/speaker-qd7vl" Nov 24 14:33:11 crc kubenswrapper[4822]: I1124 14:33:11.848707 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qd7vl" Nov 24 14:33:12 crc kubenswrapper[4822]: I1124 14:33:12.600917 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qd7vl" event={"ID":"2d55bf7d-877f-4605-9eea-b792b465e271","Type":"ContainerStarted","Data":"efa1c8d0f343a64b1d78f84c21a30abdfe8d4590cfc162a7779e48ccad874313"} Nov 24 14:33:12 crc kubenswrapper[4822]: I1124 14:33:12.601944 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qd7vl" event={"ID":"2d55bf7d-877f-4605-9eea-b792b465e271","Type":"ContainerStarted","Data":"9022cbffd12245fda54be6b7f56b921cd7c8bfe862689e4739eaa61637976050"} Nov 24 14:33:12 crc kubenswrapper[4822]: I1124 14:33:12.602046 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qd7vl" event={"ID":"2d55bf7d-877f-4605-9eea-b792b465e271","Type":"ContainerStarted","Data":"e981bb0b89329d2adeee780f953e9a2d41bf48096198001e73519cede896de3c"} Nov 24 14:33:12 crc kubenswrapper[4822]: I1124 14:33:12.602412 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qd7vl" Nov 24 14:33:12 crc kubenswrapper[4822]: I1124 14:33:12.635706 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qd7vl" podStartSLOduration=3.635682924 podStartE2EDuration="3.635682924s" podCreationTimestamp="2025-11-24 14:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:33:12.632255826 +0000 UTC m=+829.748896303" watchObservedRunningTime="2025-11-24 14:33:12.635682924 +0000 UTC m=+829.752323401" Nov 24 14:33:18 crc kubenswrapper[4822]: I1124 14:33:18.684514 4822 generic.go:334] "Generic (PLEG): container finished" podID="6ca85134-1017-4276-85a1-ac5cf4c47c0b" containerID="da31713615e4763759cf3a5350cfd8398189167ac374220f42681b2eb196e3ce" exitCode=0 Nov 24 14:33:18 crc kubenswrapper[4822]: I1124 14:33:18.684624 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerDied","Data":"da31713615e4763759cf3a5350cfd8398189167ac374220f42681b2eb196e3ce"} Nov 24 14:33:18 crc kubenswrapper[4822]: I1124 14:33:18.690794 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" event={"ID":"7144b776-5528-4a9a-9489-6e73c2608485","Type":"ContainerStarted","Data":"9e41b99b8cb122c05010cb8ed3545a88423a1caee8bb044b6ca8cb5faa6385b3"} Nov 24 14:33:18 crc kubenswrapper[4822]: I1124 14:33:18.691031 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:18 crc kubenswrapper[4822]: I1124 14:33:18.764043 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" podStartSLOduration=2.603298646 podStartE2EDuration="9.764016665s" podCreationTimestamp="2025-11-24 14:33:09 +0000 UTC" firstStartedPulling="2025-11-24 14:33:10.740781231 +0000 UTC m=+827.857421708" lastFinishedPulling="2025-11-24 14:33:17.90149925 +0000 UTC m=+835.018139727" observedRunningTime="2025-11-24 14:33:18.754705901 +0000 UTC m=+835.871346418" watchObservedRunningTime="2025-11-24 14:33:18.764016665 +0000 UTC m=+835.880657182" Nov 24 14:33:19 crc kubenswrapper[4822]: I1124 14:33:19.703580 4822 generic.go:334] "Generic (PLEG): container finished" podID="6ca85134-1017-4276-85a1-ac5cf4c47c0b" containerID="ee44b45f3442cb342d7c8b9406e16419f6d35758625fcd0b6131eddd5324e33a" exitCode=0 Nov 24 14:33:19 crc kubenswrapper[4822]: I1124 14:33:19.716577 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerDied","Data":"ee44b45f3442cb342d7c8b9406e16419f6d35758625fcd0b6131eddd5324e33a"} Nov 24 14:33:20 crc kubenswrapper[4822]: I1124 14:33:20.369897 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-dlvwp" Nov 24 14:33:20 crc kubenswrapper[4822]: I1124 14:33:20.719720 4822 generic.go:334] "Generic (PLEG): container finished" podID="6ca85134-1017-4276-85a1-ac5cf4c47c0b" containerID="6b85f111b9db37bb6357dd5b0ceecf2e13a523a52c506b67e487f5381633ae02" exitCode=0 Nov 24 14:33:20 crc kubenswrapper[4822]: I1124 14:33:20.719789 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerDied","Data":"6b85f111b9db37bb6357dd5b0ceecf2e13a523a52c506b67e487f5381633ae02"} Nov 24 14:33:21 crc kubenswrapper[4822]: I1124 14:33:21.762862 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerStarted","Data":"1a0b20bd3920d58900f5935652e2bf4ea7c6cb4b5657f5eae757ef9b9abb0103"} Nov 24 14:33:21 crc kubenswrapper[4822]: I1124 14:33:21.763291 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerStarted","Data":"6455e057924d22c83454cacbf8c73344210c25b22810509ffdc451c24e6ac694"} Nov 24 14:33:21 crc kubenswrapper[4822]: I1124 14:33:21.763310 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerStarted","Data":"dbe6bb754a4729cb9ef9424d867dc9c22da4eda39c327b1eab2701b5bcd70114"} Nov 24 14:33:21 crc kubenswrapper[4822]: I1124 14:33:21.763323 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerStarted","Data":"9d021f1ceac177ea4b1c8ceb49b93f3bfdf7dea8017aa3ff5ff5ac66aaa0265a"} Nov 24 14:33:21 crc kubenswrapper[4822]: I1124 14:33:21.763338 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerStarted","Data":"8bb9d942debec261d4957f67af697e25333cc648a1e3865dae89124acc4e2ec2"} Nov 24 14:33:22 crc kubenswrapper[4822]: I1124 14:33:22.776507 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ts9cq" event={"ID":"6ca85134-1017-4276-85a1-ac5cf4c47c0b","Type":"ContainerStarted","Data":"f4a7b4eb19218b4196eeb3cd13824c778f2685870390d22a4cc2cc3c24de8bba"} Nov 24 14:33:22 crc kubenswrapper[4822]: I1124 14:33:22.776737 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:22 crc kubenswrapper[4822]: I1124 14:33:22.805577 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ts9cq" podStartSLOduration=6.341597607 podStartE2EDuration="13.805551592s" podCreationTimestamp="2025-11-24 14:33:09 +0000 UTC" firstStartedPulling="2025-11-24 14:33:10.411991178 +0000 UTC m=+827.528631655" lastFinishedPulling="2025-11-24 14:33:17.875945163 +0000 UTC m=+834.992585640" observedRunningTime="2025-11-24 14:33:22.802191927 +0000 UTC m=+839.918832424" watchObservedRunningTime="2025-11-24 14:33:22.805551592 +0000 UTC m=+839.922192079" Nov 24 14:33:25 crc kubenswrapper[4822]: I1124 14:33:25.275534 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:25 crc kubenswrapper[4822]: I1124 14:33:25.332536 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:30 crc kubenswrapper[4822]: I1124 14:33:30.254930 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ld68" Nov 24 14:33:30 crc kubenswrapper[4822]: I1124 14:33:30.287643 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ts9cq" Nov 24 14:33:31 crc kubenswrapper[4822]: I1124 14:33:31.858617 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qd7vl" Nov 24 14:33:34 crc kubenswrapper[4822]: I1124 14:33:34.901870 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-khbqw"] Nov 24 14:33:34 crc kubenswrapper[4822]: I1124 14:33:34.904475 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-khbqw" Nov 24 14:33:34 crc kubenswrapper[4822]: I1124 14:33:34.911099 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 14:33:34 crc kubenswrapper[4822]: I1124 14:33:34.911585 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 14:33:34 crc kubenswrapper[4822]: I1124 14:33:34.911826 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-8hgkw" Nov 24 14:33:34 crc kubenswrapper[4822]: I1124 14:33:34.932865 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-khbqw"] Nov 24 14:33:34 crc kubenswrapper[4822]: I1124 14:33:34.973512 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9ww2\" (UniqueName: \"kubernetes.io/projected/daee091f-f945-43d9-a093-650e2470b9e6-kube-api-access-d9ww2\") pod \"openstack-operator-index-khbqw\" (UID: \"daee091f-f945-43d9-a093-650e2470b9e6\") " pod="openstack-operators/openstack-operator-index-khbqw" Nov 24 14:33:35 crc kubenswrapper[4822]: I1124 14:33:35.075439 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9ww2\" (UniqueName: \"kubernetes.io/projected/daee091f-f945-43d9-a093-650e2470b9e6-kube-api-access-d9ww2\") pod \"openstack-operator-index-khbqw\" (UID: \"daee091f-f945-43d9-a093-650e2470b9e6\") " pod="openstack-operators/openstack-operator-index-khbqw" Nov 24 14:33:35 crc kubenswrapper[4822]: I1124 14:33:35.123739 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9ww2\" (UniqueName: \"kubernetes.io/projected/daee091f-f945-43d9-a093-650e2470b9e6-kube-api-access-d9ww2\") pod \"openstack-operator-index-khbqw\" (UID: \"daee091f-f945-43d9-a093-650e2470b9e6\") " pod="openstack-operators/openstack-operator-index-khbqw" Nov 24 14:33:35 crc kubenswrapper[4822]: I1124 14:33:35.241608 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-khbqw" Nov 24 14:33:35 crc kubenswrapper[4822]: I1124 14:33:35.662766 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-khbqw"] Nov 24 14:33:35 crc kubenswrapper[4822]: W1124 14:33:35.672176 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddaee091f_f945_43d9_a093_650e2470b9e6.slice/crio-121a9243c6cc9913dc23dd2d1b840965cd030686878ff75ba29048d22f733a11 WatchSource:0}: Error finding container 121a9243c6cc9913dc23dd2d1b840965cd030686878ff75ba29048d22f733a11: Status 404 returned error can't find the container with id 121a9243c6cc9913dc23dd2d1b840965cd030686878ff75ba29048d22f733a11 Nov 24 14:33:35 crc kubenswrapper[4822]: I1124 14:33:35.908827 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-khbqw" event={"ID":"daee091f-f945-43d9-a093-650e2470b9e6","Type":"ContainerStarted","Data":"121a9243c6cc9913dc23dd2d1b840965cd030686878ff75ba29048d22f733a11"} Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.258771 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-khbqw"] Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.677448 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-527lx"] Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.678568 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-527lx" Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.685466 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-527lx"] Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.742430 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkjh5\" (UniqueName: \"kubernetes.io/projected/1899e20a-040e-4111-a386-e23e95850d2c-kube-api-access-qkjh5\") pod \"openstack-operator-index-527lx\" (UID: \"1899e20a-040e-4111-a386-e23e95850d2c\") " pod="openstack-operators/openstack-operator-index-527lx" Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.842982 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkjh5\" (UniqueName: \"kubernetes.io/projected/1899e20a-040e-4111-a386-e23e95850d2c-kube-api-access-qkjh5\") pod \"openstack-operator-index-527lx\" (UID: \"1899e20a-040e-4111-a386-e23e95850d2c\") " pod="openstack-operators/openstack-operator-index-527lx" Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.868383 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkjh5\" (UniqueName: \"kubernetes.io/projected/1899e20a-040e-4111-a386-e23e95850d2c-kube-api-access-qkjh5\") pod \"openstack-operator-index-527lx\" (UID: \"1899e20a-040e-4111-a386-e23e95850d2c\") " pod="openstack-operators/openstack-operator-index-527lx" Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.930709 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-khbqw" event={"ID":"daee091f-f945-43d9-a093-650e2470b9e6","Type":"ContainerStarted","Data":"313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1"} Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.930840 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-khbqw" podUID="daee091f-f945-43d9-a093-650e2470b9e6" containerName="registry-server" containerID="cri-o://313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1" gracePeriod=2 Nov 24 14:33:38 crc kubenswrapper[4822]: I1124 14:33:38.955317 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-khbqw" podStartSLOduration=2.012991187 podStartE2EDuration="4.955230902s" podCreationTimestamp="2025-11-24 14:33:34 +0000 UTC" firstStartedPulling="2025-11-24 14:33:35.675600583 +0000 UTC m=+852.792241070" lastFinishedPulling="2025-11-24 14:33:38.617840268 +0000 UTC m=+855.734480785" observedRunningTime="2025-11-24 14:33:38.953058454 +0000 UTC m=+856.069698941" watchObservedRunningTime="2025-11-24 14:33:38.955230902 +0000 UTC m=+856.071871399" Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.039917 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-527lx" Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.381606 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-khbqw" Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.453478 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9ww2\" (UniqueName: \"kubernetes.io/projected/daee091f-f945-43d9-a093-650e2470b9e6-kube-api-access-d9ww2\") pod \"daee091f-f945-43d9-a093-650e2470b9e6\" (UID: \"daee091f-f945-43d9-a093-650e2470b9e6\") " Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.461989 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daee091f-f945-43d9-a093-650e2470b9e6-kube-api-access-d9ww2" (OuterVolumeSpecName: "kube-api-access-d9ww2") pod "daee091f-f945-43d9-a093-650e2470b9e6" (UID: "daee091f-f945-43d9-a093-650e2470b9e6"). InnerVolumeSpecName "kube-api-access-d9ww2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.555875 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9ww2\" (UniqueName: \"kubernetes.io/projected/daee091f-f945-43d9-a093-650e2470b9e6-kube-api-access-d9ww2\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.581068 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-527lx"] Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.958697 4822 generic.go:334] "Generic (PLEG): container finished" podID="daee091f-f945-43d9-a093-650e2470b9e6" containerID="313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1" exitCode=0 Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.958804 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-khbqw" Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.958794 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-khbqw" event={"ID":"daee091f-f945-43d9-a093-650e2470b9e6","Type":"ContainerDied","Data":"313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1"} Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.959959 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-khbqw" event={"ID":"daee091f-f945-43d9-a093-650e2470b9e6","Type":"ContainerDied","Data":"121a9243c6cc9913dc23dd2d1b840965cd030686878ff75ba29048d22f733a11"} Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.960004 4822 scope.go:117] "RemoveContainer" containerID="313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1" Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.967399 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-527lx" event={"ID":"1899e20a-040e-4111-a386-e23e95850d2c","Type":"ContainerStarted","Data":"ab0b274ddbf5455ffcba8caf5ecb7d2d483aab3d2cf0de39434bedb9bc910975"} Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.967491 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-527lx" event={"ID":"1899e20a-040e-4111-a386-e23e95850d2c","Type":"ContainerStarted","Data":"5dfd7a7bdf91f30ae2451de467a2399518402171ee21e4e6d6f34d32387b250d"} Nov 24 14:33:39 crc kubenswrapper[4822]: I1124 14:33:39.994445 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-527lx" podStartSLOduration=1.946266805 podStartE2EDuration="1.994419346s" podCreationTimestamp="2025-11-24 14:33:38 +0000 UTC" firstStartedPulling="2025-11-24 14:33:39.592799894 +0000 UTC m=+856.709440381" lastFinishedPulling="2025-11-24 14:33:39.640952445 +0000 UTC m=+856.757592922" observedRunningTime="2025-11-24 14:33:39.994295392 +0000 UTC m=+857.110935869" watchObservedRunningTime="2025-11-24 14:33:39.994419346 +0000 UTC m=+857.111059833" Nov 24 14:33:40 crc kubenswrapper[4822]: I1124 14:33:40.003154 4822 scope.go:117] "RemoveContainer" containerID="313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1" Nov 24 14:33:40 crc kubenswrapper[4822]: E1124 14:33:40.007602 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1\": container with ID starting with 313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1 not found: ID does not exist" containerID="313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1" Nov 24 14:33:40 crc kubenswrapper[4822]: I1124 14:33:40.007643 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1"} err="failed to get container status \"313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1\": rpc error: code = NotFound desc = could not find container \"313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1\": container with ID starting with 313a8cb698b9ce03105055699e269d9b123c6386837d5e5fe7e85411dba0b8e1 not found: ID does not exist" Nov 24 14:33:40 crc kubenswrapper[4822]: I1124 14:33:40.035302 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-khbqw"] Nov 24 14:33:40 crc kubenswrapper[4822]: I1124 14:33:40.035807 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-khbqw"] Nov 24 14:33:41 crc kubenswrapper[4822]: I1124 14:33:41.730745 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daee091f-f945-43d9-a093-650e2470b9e6" path="/var/lib/kubelet/pods/daee091f-f945-43d9-a093-650e2470b9e6/volumes" Nov 24 14:33:49 crc kubenswrapper[4822]: I1124 14:33:49.040478 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-527lx" Nov 24 14:33:49 crc kubenswrapper[4822]: I1124 14:33:49.040777 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-527lx" Nov 24 14:33:49 crc kubenswrapper[4822]: I1124 14:33:49.095516 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-527lx" Nov 24 14:33:50 crc kubenswrapper[4822]: I1124 14:33:50.090066 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-527lx" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.638259 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw"] Nov 24 14:33:57 crc kubenswrapper[4822]: E1124 14:33:57.641783 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daee091f-f945-43d9-a093-650e2470b9e6" containerName="registry-server" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.641959 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="daee091f-f945-43d9-a093-650e2470b9e6" containerName="registry-server" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.642384 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="daee091f-f945-43d9-a093-650e2470b9e6" containerName="registry-server" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.645015 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.648566 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-jl2f8" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.657697 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw"] Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.676400 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-util\") pod \"f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.676781 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-bundle\") pod \"f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.676836 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwtgf\" (UniqueName: \"kubernetes.io/projected/2f49b7d5-56b4-4c58-ba64-866a06860c5a-kube-api-access-xwtgf\") pod \"f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.777752 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-bundle\") pod \"f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.777813 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwtgf\" (UniqueName: \"kubernetes.io/projected/2f49b7d5-56b4-4c58-ba64-866a06860c5a-kube-api-access-xwtgf\") pod \"f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.777895 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-util\") pod \"f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.779175 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-bundle\") pod \"f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.779810 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-util\") pod \"f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.808269 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwtgf\" (UniqueName: \"kubernetes.io/projected/2f49b7d5-56b4-4c58-ba64-866a06860c5a-kube-api-access-xwtgf\") pod \"f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:57 crc kubenswrapper[4822]: I1124 14:33:57.970165 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:33:58 crc kubenswrapper[4822]: I1124 14:33:58.539403 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw"] Nov 24 14:33:59 crc kubenswrapper[4822]: I1124 14:33:59.128391 4822 generic.go:334] "Generic (PLEG): container finished" podID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerID="b33a52c3579f2b87afec389d8de3ac95db048859442bdd42d3d73c09c85321b9" exitCode=0 Nov 24 14:33:59 crc kubenswrapper[4822]: I1124 14:33:59.128467 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" event={"ID":"2f49b7d5-56b4-4c58-ba64-866a06860c5a","Type":"ContainerDied","Data":"b33a52c3579f2b87afec389d8de3ac95db048859442bdd42d3d73c09c85321b9"} Nov 24 14:33:59 crc kubenswrapper[4822]: I1124 14:33:59.128518 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" event={"ID":"2f49b7d5-56b4-4c58-ba64-866a06860c5a","Type":"ContainerStarted","Data":"3a5121896fa838114ccee10ea4ed0771efb961c577afd8a16b6623a32d165f9f"} Nov 24 14:33:59 crc kubenswrapper[4822]: I1124 14:33:59.131178 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:34:00 crc kubenswrapper[4822]: I1124 14:34:00.139950 4822 generic.go:334] "Generic (PLEG): container finished" podID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerID="a94498f9c16c750eedfbb5efd25d0c973ce7079da735256c0acd40ff90bcc0b9" exitCode=0 Nov 24 14:34:00 crc kubenswrapper[4822]: I1124 14:34:00.140027 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" event={"ID":"2f49b7d5-56b4-4c58-ba64-866a06860c5a","Type":"ContainerDied","Data":"a94498f9c16c750eedfbb5efd25d0c973ce7079da735256c0acd40ff90bcc0b9"} Nov 24 14:34:01 crc kubenswrapper[4822]: I1124 14:34:01.159445 4822 generic.go:334] "Generic (PLEG): container finished" podID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerID="f38744c29eff02ac17c537cd818d3b434c45398395962cee81cf2e9a1c7d62a2" exitCode=0 Nov 24 14:34:01 crc kubenswrapper[4822]: I1124 14:34:01.159529 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" event={"ID":"2f49b7d5-56b4-4c58-ba64-866a06860c5a","Type":"ContainerDied","Data":"f38744c29eff02ac17c537cd818d3b434c45398395962cee81cf2e9a1c7d62a2"} Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.510585 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.553267 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwtgf\" (UniqueName: \"kubernetes.io/projected/2f49b7d5-56b4-4c58-ba64-866a06860c5a-kube-api-access-xwtgf\") pod \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.553468 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-bundle\") pod \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.553509 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-util\") pod \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\" (UID: \"2f49b7d5-56b4-4c58-ba64-866a06860c5a\") " Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.554722 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-bundle" (OuterVolumeSpecName: "bundle") pod "2f49b7d5-56b4-4c58-ba64-866a06860c5a" (UID: "2f49b7d5-56b4-4c58-ba64-866a06860c5a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.560972 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f49b7d5-56b4-4c58-ba64-866a06860c5a-kube-api-access-xwtgf" (OuterVolumeSpecName: "kube-api-access-xwtgf") pod "2f49b7d5-56b4-4c58-ba64-866a06860c5a" (UID: "2f49b7d5-56b4-4c58-ba64-866a06860c5a"). InnerVolumeSpecName "kube-api-access-xwtgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.567248 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-util" (OuterVolumeSpecName: "util") pod "2f49b7d5-56b4-4c58-ba64-866a06860c5a" (UID: "2f49b7d5-56b4-4c58-ba64-866a06860c5a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.655923 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwtgf\" (UniqueName: \"kubernetes.io/projected/2f49b7d5-56b4-4c58-ba64-866a06860c5a-kube-api-access-xwtgf\") on node \"crc\" DevicePath \"\"" Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.656838 4822 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:34:02 crc kubenswrapper[4822]: I1124 14:34:02.656856 4822 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f49b7d5-56b4-4c58-ba64-866a06860c5a-util\") on node \"crc\" DevicePath \"\"" Nov 24 14:34:03 crc kubenswrapper[4822]: I1124 14:34:03.177664 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" event={"ID":"2f49b7d5-56b4-4c58-ba64-866a06860c5a","Type":"ContainerDied","Data":"3a5121896fa838114ccee10ea4ed0771efb961c577afd8a16b6623a32d165f9f"} Nov 24 14:34:03 crc kubenswrapper[4822]: I1124 14:34:03.177721 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a5121896fa838114ccee10ea4ed0771efb961c577afd8a16b6623a32d165f9f" Nov 24 14:34:03 crc kubenswrapper[4822]: I1124 14:34:03.177845 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.110582 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8"] Nov 24 14:34:09 crc kubenswrapper[4822]: E1124 14:34:09.112817 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerName="extract" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.115196 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerName="extract" Nov 24 14:34:09 crc kubenswrapper[4822]: E1124 14:34:09.115268 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerName="util" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.115281 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerName="util" Nov 24 14:34:09 crc kubenswrapper[4822]: E1124 14:34:09.115319 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerName="pull" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.115326 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerName="pull" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.115590 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f49b7d5-56b4-4c58-ba64-866a06860c5a" containerName="extract" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.117118 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.130779 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-t5ftk" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.132035 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8"] Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.265277 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fspwh\" (UniqueName: \"kubernetes.io/projected/369ebe2a-4e3b-4131-8d92-8f8066f2fd78-kube-api-access-fspwh\") pod \"openstack-operator-controller-operator-68f4d784d9-89pr8\" (UID: \"369ebe2a-4e3b-4131-8d92-8f8066f2fd78\") " pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.367133 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fspwh\" (UniqueName: \"kubernetes.io/projected/369ebe2a-4e3b-4131-8d92-8f8066f2fd78-kube-api-access-fspwh\") pod \"openstack-operator-controller-operator-68f4d784d9-89pr8\" (UID: \"369ebe2a-4e3b-4131-8d92-8f8066f2fd78\") " pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.390074 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fspwh\" (UniqueName: \"kubernetes.io/projected/369ebe2a-4e3b-4131-8d92-8f8066f2fd78-kube-api-access-fspwh\") pod \"openstack-operator-controller-operator-68f4d784d9-89pr8\" (UID: \"369ebe2a-4e3b-4131-8d92-8f8066f2fd78\") " pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" Nov 24 14:34:09 crc kubenswrapper[4822]: I1124 14:34:09.443471 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" Nov 24 14:34:10 crc kubenswrapper[4822]: I1124 14:34:10.088629 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8"] Nov 24 14:34:10 crc kubenswrapper[4822]: I1124 14:34:10.225772 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" event={"ID":"369ebe2a-4e3b-4131-8d92-8f8066f2fd78","Type":"ContainerStarted","Data":"08c9b648c20615741b913c65c95aa98b2cfaef3dd7a068c1245d3913ad223ce4"} Nov 24 14:34:11 crc kubenswrapper[4822]: I1124 14:34:11.577940 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:34:11 crc kubenswrapper[4822]: I1124 14:34:11.578463 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:34:15 crc kubenswrapper[4822]: I1124 14:34:15.312282 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" event={"ID":"369ebe2a-4e3b-4131-8d92-8f8066f2fd78","Type":"ContainerStarted","Data":"817434df013276280e5ce8d2e515aec4551f27960af83646722b93c0fabfd8f4"} Nov 24 14:34:17 crc kubenswrapper[4822]: I1124 14:34:17.330521 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" event={"ID":"369ebe2a-4e3b-4131-8d92-8f8066f2fd78","Type":"ContainerStarted","Data":"c6902928c4dcb490fc6cf56ff5948521cf19cc120cf335326a2f68844a394c62"} Nov 24 14:34:17 crc kubenswrapper[4822]: I1124 14:34:17.331060 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" Nov 24 14:34:17 crc kubenswrapper[4822]: I1124 14:34:17.381610 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" podStartSLOduration=1.870815423 podStartE2EDuration="8.38157697s" podCreationTimestamp="2025-11-24 14:34:09 +0000 UTC" firstStartedPulling="2025-11-24 14:34:10.095938345 +0000 UTC m=+887.212578822" lastFinishedPulling="2025-11-24 14:34:16.606699892 +0000 UTC m=+893.723340369" observedRunningTime="2025-11-24 14:34:17.373528126 +0000 UTC m=+894.490168683" watchObservedRunningTime="2025-11-24 14:34:17.38157697 +0000 UTC m=+894.498217477" Nov 24 14:34:19 crc kubenswrapper[4822]: I1124 14:34:19.447629 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-68f4d784d9-89pr8" Nov 24 14:34:41 crc kubenswrapper[4822]: I1124 14:34:41.578055 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:34:41 crc kubenswrapper[4822]: I1124 14:34:41.578980 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.080859 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.083675 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.086458 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-45msn" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.096793 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.098538 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.100425 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-b4lnj" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.106892 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.113811 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.119930 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.121296 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.124342 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-bds6x" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.149619 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.151396 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.158706 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-5ncmt" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.163262 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.186440 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.198243 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.199597 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.205682 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-54c9w" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.210888 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.212312 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.218446 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-5jl5p" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.219232 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.219935 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bplpl\" (UniqueName: \"kubernetes.io/projected/70ff7dde-abd4-453c-8164-c7304292c41d-kube-api-access-bplpl\") pod \"glance-operator-controller-manager-7969689c84-xh9nm\" (UID: \"70ff7dde-abd4-453c-8164-c7304292c41d\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.220017 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5xnm\" (UniqueName: \"kubernetes.io/projected/10f3ea0d-a23e-47eb-b344-0696f8c9b1e8-kube-api-access-j5xnm\") pod \"cinder-operator-controller-manager-6498cbf48f-6f4n4\" (UID: \"10f3ea0d-a23e-47eb-b344-0696f8c9b1e8\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.220102 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx8q9\" (UniqueName: \"kubernetes.io/projected/e260fa6b-7576-4e48-8025-2c7ef2bd97f7-kube-api-access-fx8q9\") pod \"barbican-operator-controller-manager-75fb479bcc-pvs2h\" (UID: \"e260fa6b-7576-4e48-8025-2c7ef2bd97f7\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.220135 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzkbb\" (UniqueName: \"kubernetes.io/projected/3809d742-f676-41dd-aae2-d03be501e0b0-kube-api-access-gzkbb\") pod \"designate-operator-controller-manager-767ccfd65f-tvqv6\" (UID: \"3809d742-f676-41dd-aae2-d03be501e0b0\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.225007 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.312547 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.314900 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.318689 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-svvz7" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.318911 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.323924 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2rt8\" (UniqueName: \"kubernetes.io/projected/57aa63d3-2464-4cc7-a5e1-539e60ee6cca-kube-api-access-r2rt8\") pod \"heat-operator-controller-manager-56f54d6746-k6hqr\" (UID: \"57aa63d3-2464-4cc7-a5e1-539e60ee6cca\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.324085 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx8q9\" (UniqueName: \"kubernetes.io/projected/e260fa6b-7576-4e48-8025-2c7ef2bd97f7-kube-api-access-fx8q9\") pod \"barbican-operator-controller-manager-75fb479bcc-pvs2h\" (UID: \"e260fa6b-7576-4e48-8025-2c7ef2bd97f7\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.324234 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlqfq\" (UniqueName: \"kubernetes.io/projected/5047b32c-bd21-4f83-bce1-248a1ce109e9-kube-api-access-xlqfq\") pod \"horizon-operator-controller-manager-598f69df5d-4fgsn\" (UID: \"5047b32c-bd21-4f83-bce1-248a1ce109e9\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.324358 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzkbb\" (UniqueName: \"kubernetes.io/projected/3809d742-f676-41dd-aae2-d03be501e0b0-kube-api-access-gzkbb\") pod \"designate-operator-controller-manager-767ccfd65f-tvqv6\" (UID: \"3809d742-f676-41dd-aae2-d03be501e0b0\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.324488 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bplpl\" (UniqueName: \"kubernetes.io/projected/70ff7dde-abd4-453c-8164-c7304292c41d-kube-api-access-bplpl\") pod \"glance-operator-controller-manager-7969689c84-xh9nm\" (UID: \"70ff7dde-abd4-453c-8164-c7304292c41d\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.324605 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5xnm\" (UniqueName: \"kubernetes.io/projected/10f3ea0d-a23e-47eb-b344-0696f8c9b1e8-kube-api-access-j5xnm\") pod \"cinder-operator-controller-manager-6498cbf48f-6f4n4\" (UID: \"10f3ea0d-a23e-47eb-b344-0696f8c9b1e8\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.341348 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.342988 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.349607 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-5tfzb" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.366448 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.374786 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.382150 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.383671 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.385157 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx8q9\" (UniqueName: \"kubernetes.io/projected/e260fa6b-7576-4e48-8025-2c7ef2bd97f7-kube-api-access-fx8q9\") pod \"barbican-operator-controller-manager-75fb479bcc-pvs2h\" (UID: \"e260fa6b-7576-4e48-8025-2c7ef2bd97f7\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.388456 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5xnm\" (UniqueName: \"kubernetes.io/projected/10f3ea0d-a23e-47eb-b344-0696f8c9b1e8-kube-api-access-j5xnm\") pod \"cinder-operator-controller-manager-6498cbf48f-6f4n4\" (UID: \"10f3ea0d-a23e-47eb-b344-0696f8c9b1e8\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.388446 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzkbb\" (UniqueName: \"kubernetes.io/projected/3809d742-f676-41dd-aae2-d03be501e0b0-kube-api-access-gzkbb\") pod \"designate-operator-controller-manager-767ccfd65f-tvqv6\" (UID: \"3809d742-f676-41dd-aae2-d03be501e0b0\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.394373 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2zm27" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.400525 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.409028 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bplpl\" (UniqueName: \"kubernetes.io/projected/70ff7dde-abd4-453c-8164-c7304292c41d-kube-api-access-bplpl\") pod \"glance-operator-controller-manager-7969689c84-xh9nm\" (UID: \"70ff7dde-abd4-453c-8164-c7304292c41d\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.419269 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-khnxm"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.420396 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.420845 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.426809 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2mmw\" (UniqueName: \"kubernetes.io/projected/ac525cf9-4bbf-4030-a7b3-f2da129d81c8-kube-api-access-p2mmw\") pod \"keystone-operator-controller-manager-7454b96578-2vr5g\" (UID: \"ac525cf9-4bbf-4030-a7b3-f2da129d81c8\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.426877 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/37a54972-a29e-4ae1-a3f1-104bc634b3cc-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-9j9ls\" (UID: \"37a54972-a29e-4ae1-a3f1-104bc634b3cc\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.426923 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5vls\" (UniqueName: \"kubernetes.io/projected/a22d61a1-a9c4-47e9-8c5d-6593642e798b-kube-api-access-f5vls\") pod \"ironic-operator-controller-manager-99b499f4-z8mj6\" (UID: \"a22d61a1-a9c4-47e9-8c5d-6593642e798b\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.426943 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk2fw\" (UniqueName: \"kubernetes.io/projected/37a54972-a29e-4ae1-a3f1-104bc634b3cc-kube-api-access-mk2fw\") pod \"infra-operator-controller-manager-6dd8864d7c-9j9ls\" (UID: \"37a54972-a29e-4ae1-a3f1-104bc634b3cc\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.426973 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2rt8\" (UniqueName: \"kubernetes.io/projected/57aa63d3-2464-4cc7-a5e1-539e60ee6cca-kube-api-access-r2rt8\") pod \"heat-operator-controller-manager-56f54d6746-k6hqr\" (UID: \"57aa63d3-2464-4cc7-a5e1-539e60ee6cca\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.426996 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlqfq\" (UniqueName: \"kubernetes.io/projected/5047b32c-bd21-4f83-bce1-248a1ce109e9-kube-api-access-xlqfq\") pod \"horizon-operator-controller-manager-598f69df5d-4fgsn\" (UID: \"5047b32c-bd21-4f83-bce1-248a1ce109e9\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.427754 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-mtbgd" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.433700 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.437440 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-khnxm"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.448648 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.450910 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.452284 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.454791 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-fndbk" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.465816 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2rt8\" (UniqueName: \"kubernetes.io/projected/57aa63d3-2464-4cc7-a5e1-539e60ee6cca-kube-api-access-r2rt8\") pod \"heat-operator-controller-manager-56f54d6746-k6hqr\" (UID: \"57aa63d3-2464-4cc7-a5e1-539e60ee6cca\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.466952 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.468510 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.483979 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.491644 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-rwpfk" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.503003 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.511065 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlqfq\" (UniqueName: \"kubernetes.io/projected/5047b32c-bd21-4f83-bce1-248a1ce109e9-kube-api-access-xlqfq\") pod \"horizon-operator-controller-manager-598f69df5d-4fgsn\" (UID: \"5047b32c-bd21-4f83-bce1-248a1ce109e9\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.518537 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.525864 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.528578 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5vls\" (UniqueName: \"kubernetes.io/projected/a22d61a1-a9c4-47e9-8c5d-6593642e798b-kube-api-access-f5vls\") pod \"ironic-operator-controller-manager-99b499f4-z8mj6\" (UID: \"a22d61a1-a9c4-47e9-8c5d-6593642e798b\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.528615 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk2fw\" (UniqueName: \"kubernetes.io/projected/37a54972-a29e-4ae1-a3f1-104bc634b3cc-kube-api-access-mk2fw\") pod \"infra-operator-controller-manager-6dd8864d7c-9j9ls\" (UID: \"37a54972-a29e-4ae1-a3f1-104bc634b3cc\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.528664 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2mmw\" (UniqueName: \"kubernetes.io/projected/ac525cf9-4bbf-4030-a7b3-f2da129d81c8-kube-api-access-p2mmw\") pod \"keystone-operator-controller-manager-7454b96578-2vr5g\" (UID: \"ac525cf9-4bbf-4030-a7b3-f2da129d81c8\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.528691 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snllq\" (UniqueName: \"kubernetes.io/projected/8939f780-02b5-4022-80cb-a6fd247aa7f8-kube-api-access-snllq\") pod \"manila-operator-controller-manager-58f887965d-khnxm\" (UID: \"8939f780-02b5-4022-80cb-a6fd247aa7f8\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.528715 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khgls\" (UniqueName: \"kubernetes.io/projected/3e001a57-d6f1-4d8a-9d51-f347063a55b5-kube-api-access-khgls\") pod \"mariadb-operator-controller-manager-54b5986bb8-jkq8m\" (UID: \"3e001a57-d6f1-4d8a-9d51-f347063a55b5\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.528760 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/37a54972-a29e-4ae1-a3f1-104bc634b3cc-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-9j9ls\" (UID: \"37a54972-a29e-4ae1-a3f1-104bc634b3cc\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.528799 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwd7r\" (UniqueName: \"kubernetes.io/projected/80740002-5d02-4aa7-b6fa-3de642266f21-kube-api-access-wwd7r\") pod \"neutron-operator-controller-manager-78bd47f458-bzj92\" (UID: \"80740002-5d02-4aa7-b6fa-3de642266f21\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" Nov 24 14:34:55 crc kubenswrapper[4822]: E1124 14:34:55.529473 4822 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 14:34:55 crc kubenswrapper[4822]: E1124 14:34:55.529525 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37a54972-a29e-4ae1-a3f1-104bc634b3cc-cert podName:37a54972-a29e-4ae1-a3f1-104bc634b3cc nodeName:}" failed. No retries permitted until 2025-11-24 14:34:56.029505115 +0000 UTC m=+933.146145592 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/37a54972-a29e-4ae1-a3f1-104bc634b3cc-cert") pod "infra-operator-controller-manager-6dd8864d7c-9j9ls" (UID: "37a54972-a29e-4ae1-a3f1-104bc634b3cc") : secret "infra-operator-webhook-server-cert" not found Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.540238 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.556643 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk2fw\" (UniqueName: \"kubernetes.io/projected/37a54972-a29e-4ae1-a3f1-104bc634b3cc-kube-api-access-mk2fw\") pod \"infra-operator-controller-manager-6dd8864d7c-9j9ls\" (UID: \"37a54972-a29e-4ae1-a3f1-104bc634b3cc\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.557260 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5vls\" (UniqueName: \"kubernetes.io/projected/a22d61a1-a9c4-47e9-8c5d-6593642e798b-kube-api-access-f5vls\") pod \"ironic-operator-controller-manager-99b499f4-z8mj6\" (UID: \"a22d61a1-a9c4-47e9-8c5d-6593642e798b\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.563661 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2mmw\" (UniqueName: \"kubernetes.io/projected/ac525cf9-4bbf-4030-a7b3-f2da129d81c8-kube-api-access-p2mmw\") pod \"keystone-operator-controller-manager-7454b96578-2vr5g\" (UID: \"ac525cf9-4bbf-4030-a7b3-f2da129d81c8\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.580879 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.581008 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.586418 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-fd4v9" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.604263 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.612947 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.614602 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.621438 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-7xhxv" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.634930 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.635983 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwz27\" (UniqueName: \"kubernetes.io/projected/fd0cee72-09cc-449b-a6f7-fb951f0b87ba-kube-api-access-vwz27\") pod \"nova-operator-controller-manager-cfbb9c588-xd82t\" (UID: \"fd0cee72-09cc-449b-a6f7-fb951f0b87ba\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.641370 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snllq\" (UniqueName: \"kubernetes.io/projected/8939f780-02b5-4022-80cb-a6fd247aa7f8-kube-api-access-snllq\") pod \"manila-operator-controller-manager-58f887965d-khnxm\" (UID: \"8939f780-02b5-4022-80cb-a6fd247aa7f8\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.641408 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khgls\" (UniqueName: \"kubernetes.io/projected/3e001a57-d6f1-4d8a-9d51-f347063a55b5-kube-api-access-khgls\") pod \"mariadb-operator-controller-manager-54b5986bb8-jkq8m\" (UID: \"3e001a57-d6f1-4d8a-9d51-f347063a55b5\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.641502 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.641523 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwd7r\" (UniqueName: \"kubernetes.io/projected/80740002-5d02-4aa7-b6fa-3de642266f21-kube-api-access-wwd7r\") pod \"neutron-operator-controller-manager-78bd47f458-bzj92\" (UID: \"80740002-5d02-4aa7-b6fa-3de642266f21\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.646525 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.657798 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-lxkzf" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.658080 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.678856 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.680632 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.687459 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwd7r\" (UniqueName: \"kubernetes.io/projected/80740002-5d02-4aa7-b6fa-3de642266f21-kube-api-access-wwd7r\") pod \"neutron-operator-controller-manager-78bd47f458-bzj92\" (UID: \"80740002-5d02-4aa7-b6fa-3de642266f21\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.689183 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khgls\" (UniqueName: \"kubernetes.io/projected/3e001a57-d6f1-4d8a-9d51-f347063a55b5-kube-api-access-khgls\") pod \"mariadb-operator-controller-manager-54b5986bb8-jkq8m\" (UID: \"3e001a57-d6f1-4d8a-9d51-f347063a55b5\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.703861 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-8n22g" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.704628 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snllq\" (UniqueName: \"kubernetes.io/projected/8939f780-02b5-4022-80cb-a6fd247aa7f8-kube-api-access-snllq\") pod \"manila-operator-controller-manager-58f887965d-khnxm\" (UID: \"8939f780-02b5-4022-80cb-a6fd247aa7f8\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.704694 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.754827 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgvsh\" (UniqueName: \"kubernetes.io/projected/b5737c88-eb91-432d-a620-734337051260-kube-api-access-tgvsh\") pod \"octavia-operator-controller-manager-54cfbf4c7d-mf4hq\" (UID: \"b5737c88-eb91-432d-a620-734337051260\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.754966 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ljw6\" (UniqueName: \"kubernetes.io/projected/dbffd719-e3e2-4728-9301-da9efe8eb57b-kube-api-access-7ljw6\") pod \"ovn-operator-controller-manager-54fc5f65b7-lsbgj\" (UID: \"dbffd719-e3e2-4728-9301-da9efe8eb57b\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.755125 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b01f0f96-a40f-407a-a610-bd2e945a3e49-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc\" (UID: \"b01f0f96-a40f-407a-a610-bd2e945a3e49\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.755232 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwz27\" (UniqueName: \"kubernetes.io/projected/fd0cee72-09cc-449b-a6f7-fb951f0b87ba-kube-api-access-vwz27\") pod \"nova-operator-controller-manager-cfbb9c588-xd82t\" (UID: \"fd0cee72-09cc-449b-a6f7-fb951f0b87ba\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.755290 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w55h\" (UniqueName: \"kubernetes.io/projected/b01f0f96-a40f-407a-a610-bd2e945a3e49-kube-api-access-8w55h\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc\" (UID: \"b01f0f96-a40f-407a-a610-bd2e945a3e49\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.792245 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.805062 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.805934 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwz27\" (UniqueName: \"kubernetes.io/projected/fd0cee72-09cc-449b-a6f7-fb951f0b87ba-kube-api-access-vwz27\") pod \"nova-operator-controller-manager-cfbb9c588-xd82t\" (UID: \"fd0cee72-09cc-449b-a6f7-fb951f0b87ba\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.807320 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.807434 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.811625 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.826636 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-6sh6g" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.826714 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.828959 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-x2ftw" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.833342 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.834645 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.842680 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.844982 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-wsnnj" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.847096 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.850104 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.886591 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b01f0f96-a40f-407a-a610-bd2e945a3e49-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc\" (UID: \"b01f0f96-a40f-407a-a610-bd2e945a3e49\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.886780 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w55h\" (UniqueName: \"kubernetes.io/projected/b01f0f96-a40f-407a-a610-bd2e945a3e49-kube-api-access-8w55h\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc\" (UID: \"b01f0f96-a40f-407a-a610-bd2e945a3e49\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.886932 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgvsh\" (UniqueName: \"kubernetes.io/projected/b5737c88-eb91-432d-a620-734337051260-kube-api-access-tgvsh\") pod \"octavia-operator-controller-manager-54cfbf4c7d-mf4hq\" (UID: \"b5737c88-eb91-432d-a620-734337051260\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.886991 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ljw6\" (UniqueName: \"kubernetes.io/projected/dbffd719-e3e2-4728-9301-da9efe8eb57b-kube-api-access-7ljw6\") pod \"ovn-operator-controller-manager-54fc5f65b7-lsbgj\" (UID: \"dbffd719-e3e2-4728-9301-da9efe8eb57b\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.887056 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9xzk\" (UniqueName: \"kubernetes.io/projected/d758b1f7-956b-4881-8a61-a81e55fdffcd-kube-api-access-q9xzk\") pod \"placement-operator-controller-manager-5b797b8dff-45wdc\" (UID: \"d758b1f7-956b-4881-8a61-a81e55fdffcd\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" Nov 24 14:34:55 crc kubenswrapper[4822]: E1124 14:34:55.887349 4822 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 14:34:55 crc kubenswrapper[4822]: E1124 14:34:55.887449 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b01f0f96-a40f-407a-a610-bd2e945a3e49-cert podName:b01f0f96-a40f-407a-a610-bd2e945a3e49 nodeName:}" failed. No retries permitted until 2025-11-24 14:34:56.387417523 +0000 UTC m=+933.504058000 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b01f0f96-a40f-407a-a610-bd2e945a3e49-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" (UID: "b01f0f96-a40f-407a-a610-bd2e945a3e49") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.891642 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k"] Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.917437 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.928467 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.942394 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.968240 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgvsh\" (UniqueName: \"kubernetes.io/projected/b5737c88-eb91-432d-a620-734337051260-kube-api-access-tgvsh\") pod \"octavia-operator-controller-manager-54cfbf4c7d-mf4hq\" (UID: \"b5737c88-eb91-432d-a620-734337051260\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.969486 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w55h\" (UniqueName: \"kubernetes.io/projected/b01f0f96-a40f-407a-a610-bd2e945a3e49-kube-api-access-8w55h\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc\" (UID: \"b01f0f96-a40f-407a-a610-bd2e945a3e49\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.993302 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9xzk\" (UniqueName: \"kubernetes.io/projected/d758b1f7-956b-4881-8a61-a81e55fdffcd-kube-api-access-q9xzk\") pod \"placement-operator-controller-manager-5b797b8dff-45wdc\" (UID: \"d758b1f7-956b-4881-8a61-a81e55fdffcd\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.993501 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82p7d\" (UniqueName: \"kubernetes.io/projected/38159b16-c3af-4bcd-8734-88cfef861663-kube-api-access-82p7d\") pod \"swift-operator-controller-manager-d656998f4-wtv4t\" (UID: \"38159b16-c3af-4bcd-8734-88cfef861663\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" Nov 24 14:34:55 crc kubenswrapper[4822]: I1124 14:34:55.993663 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6czgw\" (UniqueName: \"kubernetes.io/projected/253630d1-2b9c-4bc4-a7fd-357c66d00f9e-kube-api-access-6czgw\") pod \"telemetry-operator-controller-manager-54d7678447-dx47k\" (UID: \"253630d1-2b9c-4bc4-a7fd-357c66d00f9e\") " pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.014323 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.036660 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ljw6\" (UniqueName: \"kubernetes.io/projected/dbffd719-e3e2-4728-9301-da9efe8eb57b-kube-api-access-7ljw6\") pod \"ovn-operator-controller-manager-54fc5f65b7-lsbgj\" (UID: \"dbffd719-e3e2-4728-9301-da9efe8eb57b\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.058805 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9xzk\" (UniqueName: \"kubernetes.io/projected/d758b1f7-956b-4881-8a61-a81e55fdffcd-kube-api-access-q9xzk\") pod \"placement-operator-controller-manager-5b797b8dff-45wdc\" (UID: \"d758b1f7-956b-4881-8a61-a81e55fdffcd\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.061432 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.063230 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.068077 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4dxw6" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.072865 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.085796 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.095691 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82p7d\" (UniqueName: \"kubernetes.io/projected/38159b16-c3af-4bcd-8734-88cfef861663-kube-api-access-82p7d\") pod \"swift-operator-controller-manager-d656998f4-wtv4t\" (UID: \"38159b16-c3af-4bcd-8734-88cfef861663\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.096088 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/37a54972-a29e-4ae1-a3f1-104bc634b3cc-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-9j9ls\" (UID: \"37a54972-a29e-4ae1-a3f1-104bc634b3cc\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.096226 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6czgw\" (UniqueName: \"kubernetes.io/projected/253630d1-2b9c-4bc4-a7fd-357c66d00f9e-kube-api-access-6czgw\") pod \"telemetry-operator-controller-manager-54d7678447-dx47k\" (UID: \"253630d1-2b9c-4bc4-a7fd-357c66d00f9e\") " pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.100377 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.103859 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/37a54972-a29e-4ae1-a3f1-104bc634b3cc-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-9j9ls\" (UID: \"37a54972-a29e-4ae1-a3f1-104bc634b3cc\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.106643 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.109193 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.114861 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-kpbmc" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.117893 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6czgw\" (UniqueName: \"kubernetes.io/projected/253630d1-2b9c-4bc4-a7fd-357c66d00f9e-kube-api-access-6czgw\") pod \"telemetry-operator-controller-manager-54d7678447-dx47k\" (UID: \"253630d1-2b9c-4bc4-a7fd-357c66d00f9e\") " pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.128643 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.135650 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82p7d\" (UniqueName: \"kubernetes.io/projected/38159b16-c3af-4bcd-8734-88cfef861663-kube-api-access-82p7d\") pod \"swift-operator-controller-manager-d656998f4-wtv4t\" (UID: \"38159b16-c3af-4bcd-8734-88cfef861663\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.163255 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.165555 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.167742 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.167961 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-gw4lh" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.179334 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.199057 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwd8r\" (UniqueName: \"kubernetes.io/projected/97913a25-5fcf-4541-b928-191ffe821e93-kube-api-access-mwd8r\") pod \"test-operator-controller-manager-b4c496f69-2p2qp\" (UID: \"97913a25-5fcf-4541-b928-191ffe821e93\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.199165 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wp8z\" (UniqueName: \"kubernetes.io/projected/980ad0a7-70cd-4f77-bc29-119868034087-kube-api-access-7wp8z\") pod \"watcher-operator-controller-manager-8c6448b9f-dc279\" (UID: \"980ad0a7-70cd-4f77-bc29-119868034087\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.212177 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.213443 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.216362 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.217494 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-gfzmc" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.233120 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.249697 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.281582 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.301149 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdn7g\" (UniqueName: \"kubernetes.io/projected/4e5bcbaf-c685-4047-a37a-3c058f8e1cbc-kube-api-access-qdn7g\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-98xcn\" (UID: \"4e5bcbaf-c685-4047-a37a-3c058f8e1cbc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.301232 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn9bv\" (UniqueName: \"kubernetes.io/projected/a648d53c-be72-46ed-8308-79565a88bd4c-kube-api-access-tn9bv\") pod \"openstack-operator-controller-manager-7b58687767-tfthk\" (UID: \"a648d53c-be72-46ed-8308-79565a88bd4c\") " pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.301280 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwd8r\" (UniqueName: \"kubernetes.io/projected/97913a25-5fcf-4541-b928-191ffe821e93-kube-api-access-mwd8r\") pod \"test-operator-controller-manager-b4c496f69-2p2qp\" (UID: \"97913a25-5fcf-4541-b928-191ffe821e93\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.301330 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a648d53c-be72-46ed-8308-79565a88bd4c-cert\") pod \"openstack-operator-controller-manager-7b58687767-tfthk\" (UID: \"a648d53c-be72-46ed-8308-79565a88bd4c\") " pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.301388 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wp8z\" (UniqueName: \"kubernetes.io/projected/980ad0a7-70cd-4f77-bc29-119868034087-kube-api-access-7wp8z\") pod \"watcher-operator-controller-manager-8c6448b9f-dc279\" (UID: \"980ad0a7-70cd-4f77-bc29-119868034087\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.323155 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wp8z\" (UniqueName: \"kubernetes.io/projected/980ad0a7-70cd-4f77-bc29-119868034087-kube-api-access-7wp8z\") pod \"watcher-operator-controller-manager-8c6448b9f-dc279\" (UID: \"980ad0a7-70cd-4f77-bc29-119868034087\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.327605 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwd8r\" (UniqueName: \"kubernetes.io/projected/97913a25-5fcf-4541-b928-191ffe821e93-kube-api-access-mwd8r\") pod \"test-operator-controller-manager-b4c496f69-2p2qp\" (UID: \"97913a25-5fcf-4541-b928-191ffe821e93\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.360340 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.403999 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdn7g\" (UniqueName: \"kubernetes.io/projected/4e5bcbaf-c685-4047-a37a-3c058f8e1cbc-kube-api-access-qdn7g\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-98xcn\" (UID: \"4e5bcbaf-c685-4047-a37a-3c058f8e1cbc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.404050 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn9bv\" (UniqueName: \"kubernetes.io/projected/a648d53c-be72-46ed-8308-79565a88bd4c-kube-api-access-tn9bv\") pod \"openstack-operator-controller-manager-7b58687767-tfthk\" (UID: \"a648d53c-be72-46ed-8308-79565a88bd4c\") " pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.404109 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a648d53c-be72-46ed-8308-79565a88bd4c-cert\") pod \"openstack-operator-controller-manager-7b58687767-tfthk\" (UID: \"a648d53c-be72-46ed-8308-79565a88bd4c\") " pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.404133 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b01f0f96-a40f-407a-a610-bd2e945a3e49-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc\" (UID: \"b01f0f96-a40f-407a-a610-bd2e945a3e49\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:34:56 crc kubenswrapper[4822]: E1124 14:34:56.406065 4822 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 14:34:56 crc kubenswrapper[4822]: E1124 14:34:56.406153 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a648d53c-be72-46ed-8308-79565a88bd4c-cert podName:a648d53c-be72-46ed-8308-79565a88bd4c nodeName:}" failed. No retries permitted until 2025-11-24 14:34:56.90612769 +0000 UTC m=+934.022768167 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a648d53c-be72-46ed-8308-79565a88bd4c-cert") pod "openstack-operator-controller-manager-7b58687767-tfthk" (UID: "a648d53c-be72-46ed-8308-79565a88bd4c") : secret "webhook-server-cert" not found Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.417009 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b01f0f96-a40f-407a-a610-bd2e945a3e49-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc\" (UID: \"b01f0f96-a40f-407a-a610-bd2e945a3e49\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.430948 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn9bv\" (UniqueName: \"kubernetes.io/projected/a648d53c-be72-46ed-8308-79565a88bd4c-kube-api-access-tn9bv\") pod \"openstack-operator-controller-manager-7b58687767-tfthk\" (UID: \"a648d53c-be72-46ed-8308-79565a88bd4c\") " pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.460372 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdn7g\" (UniqueName: \"kubernetes.io/projected/4e5bcbaf-c685-4047-a37a-3c058f8e1cbc-kube-api-access-qdn7g\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-98xcn\" (UID: \"4e5bcbaf-c685-4047-a37a-3c058f8e1cbc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.502337 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.506866 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.509922 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4"] Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.524493 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.627983 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.675407 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.794888 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" event={"ID":"10f3ea0d-a23e-47eb-b344-0696f8c9b1e8","Type":"ContainerStarted","Data":"0ba8e8e440a7e6592aeb017662a2f4d34f2ff5d5e69ef478b0cc005f2cf24e70"} Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.796547 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" event={"ID":"e260fa6b-7576-4e48-8025-2c7ef2bd97f7","Type":"ContainerStarted","Data":"9c95acf758df5c2c1fb959f5cd2f1909aad0c81d59c457e6bbebac763582682b"} Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.919451 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a648d53c-be72-46ed-8308-79565a88bd4c-cert\") pod \"openstack-operator-controller-manager-7b58687767-tfthk\" (UID: \"a648d53c-be72-46ed-8308-79565a88bd4c\") " pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.930140 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a648d53c-be72-46ed-8308-79565a88bd4c-cert\") pod \"openstack-operator-controller-manager-7b58687767-tfthk\" (UID: \"a648d53c-be72-46ed-8308-79565a88bd4c\") " pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.978875 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm"] Nov 24 14:34:56 crc kubenswrapper[4822]: W1124 14:34:56.994187 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5047b32c_bd21_4f83_bce1_248a1ce109e9.slice/crio-6e88b66ce21245edb8beec00cc5132f14816a5c35daca3ad56e5804c66dfbffa WatchSource:0}: Error finding container 6e88b66ce21245edb8beec00cc5132f14816a5c35daca3ad56e5804c66dfbffa: Status 404 returned error can't find the container with id 6e88b66ce21245edb8beec00cc5132f14816a5c35daca3ad56e5804c66dfbffa Nov 24 14:34:56 crc kubenswrapper[4822]: W1124 14:34:56.996271 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57aa63d3_2464_4cc7_a5e1_539e60ee6cca.slice/crio-91f253e5851f299ecb34d4438f019b61defe30f04ae561467974f050d70920e5 WatchSource:0}: Error finding container 91f253e5851f299ecb34d4438f019b61defe30f04ae561467974f050d70920e5: Status 404 returned error can't find the container with id 91f253e5851f299ecb34d4438f019b61defe30f04ae561467974f050d70920e5 Nov 24 14:34:56 crc kubenswrapper[4822]: I1124 14:34:56.997030 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.003121 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr"] Nov 24 14:34:57 crc kubenswrapper[4822]: W1124 14:34:57.155612 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3809d742_f676_41dd_aae2_d03be501e0b0.slice/crio-4fca4821f406e72ba26f6af218bb000f2b9508a6b7e3e77e11990858a9206b96 WatchSource:0}: Error finding container 4fca4821f406e72ba26f6af218bb000f2b9508a6b7e3e77e11990858a9206b96: Status 404 returned error can't find the container with id 4fca4821f406e72ba26f6af218bb000f2b9508a6b7e3e77e11990858a9206b96 Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.160307 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.160387 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.438690 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.479982 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.503044 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t"] Nov 24 14:34:57 crc kubenswrapper[4822]: W1124 14:34:57.511657 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbffd719_e3e2_4728_9301_da9efe8eb57b.slice/crio-19468220dbfe948237039853f923f68f882fc03839352af0e567ad212bffd2b0 WatchSource:0}: Error finding container 19468220dbfe948237039853f923f68f882fc03839352af0e567ad212bffd2b0: Status 404 returned error can't find the container with id 19468220dbfe948237039853f923f68f882fc03839352af0e567ad212bffd2b0 Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.526652 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.534043 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.541112 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.552345 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.560922 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.566772 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-khnxm"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.572524 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m"] Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.578473 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p2mmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7454b96578-2vr5g_openstack-operators(ac525cf9-4bbf-4030-a7b3-f2da129d81c8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.583743 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-khgls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-54b5986bb8-jkq8m_openstack-operators(3e001a57-d6f1-4d8a-9d51-f347063a55b5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.588701 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7wp8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-dc279_openstack-operators(980ad0a7-70cd-4f77-bc29-119868034087): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.595393 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f5vls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-99b499f4-z8mj6_openstack-operators(a22d61a1-a9c4-47e9-8c5d-6593642e798b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 14:34:57 crc kubenswrapper[4822]: W1124 14:34:57.596454 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97913a25_5fcf_4541_b928_191ffe821e93.slice/crio-4b0d9acfbaad9c40d22d3d0a2aa68b827952dc99638530ffd7fbbb2469403cbb WatchSource:0}: Error finding container 4b0d9acfbaad9c40d22d3d0a2aa68b827952dc99638530ffd7fbbb2469403cbb: Status 404 returned error can't find the container with id 4b0d9acfbaad9c40d22d3d0a2aa68b827952dc99638530ffd7fbbb2469403cbb Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.597773 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t"] Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.614764 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mwd8r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-2p2qp_openstack-operators(97913a25-5fcf-4541-b928-191ffe821e93): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.627987 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279"] Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.634061 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.73:5001/openstack-k8s-operators/telemetry-operator:57284e36a25ee63dd9ec7bcaee4ba1f50bec5c34,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6czgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-54d7678447-dx47k_openstack-operators(253630d1-2b9c-4bc4-a7fd-357c66d00f9e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.638745 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.649987 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.748716 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn"] Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.752634 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc"] Nov 24 14:34:57 crc kubenswrapper[4822]: W1124 14:34:57.766686 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb01f0f96_a40f_407a_a610_bd2e945a3e49.slice/crio-4dc6fcd6ebdc03525b7bf98c3cc57edf1e3cd0fcc86c9c59980588abdcc998b7 WatchSource:0}: Error finding container 4dc6fcd6ebdc03525b7bf98c3cc57edf1e3cd0fcc86c9c59980588abdcc998b7: Status 404 returned error can't find the container with id 4dc6fcd6ebdc03525b7bf98c3cc57edf1e3cd0fcc86c9c59980588abdcc998b7 Nov 24 14:34:57 crc kubenswrapper[4822]: W1124 14:34:57.795410 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e5bcbaf_c685_4047_a37a_3c058f8e1cbc.slice/crio-dd47f0006585dcc0c6c15e9aacd0e150ebac1370d5455b72f33f6dc0c8d07090 WatchSource:0}: Error finding container dd47f0006585dcc0c6c15e9aacd0e150ebac1370d5455b72f33f6dc0c8d07090: Status 404 returned error can't find the container with id dd47f0006585dcc0c6c15e9aacd0e150ebac1370d5455b72f33f6dc0c8d07090 Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.814868 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk"] Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.820273 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qdn7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-98xcn_openstack-operators(4e5bcbaf-c685-4047-a37a-3c058f8e1cbc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.823034 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" podUID="4e5bcbaf-c685-4047-a37a-3c058f8e1cbc" Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.827871 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" event={"ID":"80740002-5d02-4aa7-b6fa-3de642266f21","Type":"ContainerStarted","Data":"5e4cc961bfffce3205f0e70d9b5ef4b91be64b923abb94bf36cb9b5b01ec81a2"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.831797 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" event={"ID":"57aa63d3-2464-4cc7-a5e1-539e60ee6cca","Type":"ContainerStarted","Data":"91f253e5851f299ecb34d4438f019b61defe30f04ae561467974f050d70920e5"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.834273 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" event={"ID":"b01f0f96-a40f-407a-a610-bd2e945a3e49","Type":"ContainerStarted","Data":"4dc6fcd6ebdc03525b7bf98c3cc57edf1e3cd0fcc86c9c59980588abdcc998b7"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.846118 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" event={"ID":"fd0cee72-09cc-449b-a6f7-fb951f0b87ba","Type":"ContainerStarted","Data":"87a6a9d500670603bcd45c2a76419ee387080a65f46600e9390756b1eb1ef19c"} Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.847136 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" podUID="ac525cf9-4bbf-4030-a7b3-f2da129d81c8" Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.851707 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" event={"ID":"3809d742-f676-41dd-aae2-d03be501e0b0","Type":"ContainerStarted","Data":"4fca4821f406e72ba26f6af218bb000f2b9508a6b7e3e77e11990858a9206b96"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.855918 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" event={"ID":"b5737c88-eb91-432d-a620-734337051260","Type":"ContainerStarted","Data":"a53114b6df6a11a0f6d53534538604fa549de826916b546c0f3915b242029ab1"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.858951 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" event={"ID":"38159b16-c3af-4bcd-8734-88cfef861663","Type":"ContainerStarted","Data":"b7501e5fbd06422e6f538ca1b2638795ebea73397097c0f550b339478753a78e"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.864966 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" event={"ID":"4e5bcbaf-c685-4047-a37a-3c058f8e1cbc","Type":"ContainerStarted","Data":"dd47f0006585dcc0c6c15e9aacd0e150ebac1370d5455b72f33f6dc0c8d07090"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.868449 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" event={"ID":"37a54972-a29e-4ae1-a3f1-104bc634b3cc","Type":"ContainerStarted","Data":"454bfe7aec1716d7e58e337a72445313cf0433655bb85f209dc8a341c4cb9042"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.870321 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" event={"ID":"97913a25-5fcf-4541-b928-191ffe821e93","Type":"ContainerStarted","Data":"4b0d9acfbaad9c40d22d3d0a2aa68b827952dc99638530ffd7fbbb2469403cbb"} Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.871137 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" podUID="4e5bcbaf-c685-4047-a37a-3c058f8e1cbc" Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.876831 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" event={"ID":"8939f780-02b5-4022-80cb-a6fd247aa7f8","Type":"ContainerStarted","Data":"ad31c9fa2f5733d19fdf79869a2b60b205654c7c9787aa37b1823097a46fd149"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.878784 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" event={"ID":"5047b32c-bd21-4f83-bce1-248a1ce109e9","Type":"ContainerStarted","Data":"6e88b66ce21245edb8beec00cc5132f14816a5c35daca3ad56e5804c66dfbffa"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.881341 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" event={"ID":"d758b1f7-956b-4881-8a61-a81e55fdffcd","Type":"ContainerStarted","Data":"f786a3839441468d920305eee0f82cae1febf939ad4222884475cebc39986571"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.892562 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" event={"ID":"dbffd719-e3e2-4728-9301-da9efe8eb57b","Type":"ContainerStarted","Data":"19468220dbfe948237039853f923f68f882fc03839352af0e567ad212bffd2b0"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.894637 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" event={"ID":"70ff7dde-abd4-453c-8164-c7304292c41d","Type":"ContainerStarted","Data":"f31bcc46eff62be45d3655765fea8d34b36099d8cf47d5dc6b007e2e86e17e66"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.899553 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" event={"ID":"980ad0a7-70cd-4f77-bc29-119868034087","Type":"ContainerStarted","Data":"a5d139a6384860cfa693aa54bbb5b6c676e80c6f8876bde0353953b2a3d24a76"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.901006 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" event={"ID":"a22d61a1-a9c4-47e9-8c5d-6593642e798b","Type":"ContainerStarted","Data":"33fe875bf07785718a2417b67f17409faae794999af38419c42bf9d634023ce7"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.904442 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" event={"ID":"3e001a57-d6f1-4d8a-9d51-f347063a55b5","Type":"ContainerStarted","Data":"a98f0c31feaa6feb4f3f731068be65166f04799ca6044f1e996a75ab988cac16"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.908408 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" event={"ID":"ac525cf9-4bbf-4030-a7b3-f2da129d81c8","Type":"ContainerStarted","Data":"feafffe426b5852d5facad276774c9d7900e561ae2395abf36cf81118d6d1cc3"} Nov 24 14:34:57 crc kubenswrapper[4822]: I1124 14:34:57.912256 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" event={"ID":"253630d1-2b9c-4bc4-a7fd-357c66d00f9e","Type":"ContainerStarted","Data":"194a15144fb542250e4ecb495a7a45eadd75032fbb509702d7f870cac9e607af"} Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.912333 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" podUID="ac525cf9-4bbf-4030-a7b3-f2da129d81c8" Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.955159 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" podUID="980ad0a7-70cd-4f77-bc29-119868034087" Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.968683 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" podUID="3e001a57-d6f1-4d8a-9d51-f347063a55b5" Nov 24 14:34:57 crc kubenswrapper[4822]: E1124 14:34:57.968891 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" podUID="a22d61a1-a9c4-47e9-8c5d-6593642e798b" Nov 24 14:34:58 crc kubenswrapper[4822]: E1124 14:34:58.047699 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" podUID="97913a25-5fcf-4541-b928-191ffe821e93" Nov 24 14:34:58 crc kubenswrapper[4822]: E1124 14:34:58.048368 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" podUID="253630d1-2b9c-4bc4-a7fd-357c66d00f9e" Nov 24 14:34:58 crc kubenswrapper[4822]: I1124 14:34:58.927032 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" event={"ID":"97913a25-5fcf-4541-b928-191ffe821e93","Type":"ContainerStarted","Data":"193e311491a8ba4478a7592a4d74d0ea36a949084f6600ac83a9d69a409fb37e"} Nov 24 14:34:58 crc kubenswrapper[4822]: E1124 14:34:58.933465 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" podUID="97913a25-5fcf-4541-b928-191ffe821e93" Nov 24 14:34:58 crc kubenswrapper[4822]: I1124 14:34:58.971734 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" event={"ID":"253630d1-2b9c-4bc4-a7fd-357c66d00f9e","Type":"ContainerStarted","Data":"0114472b255e9f5004af555b017a3aceb9ceb2696443b2974c400fe8fcf7b9fd"} Nov 24 14:34:58 crc kubenswrapper[4822]: E1124 14:34:58.976666 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.73:5001/openstack-k8s-operators/telemetry-operator:57284e36a25ee63dd9ec7bcaee4ba1f50bec5c34\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" podUID="253630d1-2b9c-4bc4-a7fd-357c66d00f9e" Nov 24 14:34:58 crc kubenswrapper[4822]: I1124 14:34:58.977181 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" event={"ID":"980ad0a7-70cd-4f77-bc29-119868034087","Type":"ContainerStarted","Data":"c36bb14005c75ad181c6dd819fed3ef69b474c3cbe4f816ad9a8b88975f00bc4"} Nov 24 14:34:58 crc kubenswrapper[4822]: E1124 14:34:58.982871 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" podUID="980ad0a7-70cd-4f77-bc29-119868034087" Nov 24 14:34:58 crc kubenswrapper[4822]: I1124 14:34:58.997805 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" event={"ID":"a22d61a1-a9c4-47e9-8c5d-6593642e798b","Type":"ContainerStarted","Data":"74d43cf361da015f19b4493f00624dfe912729417f7e7e6d1621487e550561a3"} Nov 24 14:34:59 crc kubenswrapper[4822]: E1124 14:34:59.012983 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" podUID="a22d61a1-a9c4-47e9-8c5d-6593642e798b" Nov 24 14:34:59 crc kubenswrapper[4822]: I1124 14:34:59.019074 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" event={"ID":"3e001a57-d6f1-4d8a-9d51-f347063a55b5","Type":"ContainerStarted","Data":"ac3ec206895e19fede5f0ca1e85a850a342b07cd271426a986cfaccb96509e01"} Nov 24 14:34:59 crc kubenswrapper[4822]: E1124 14:34:59.036603 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" podUID="3e001a57-d6f1-4d8a-9d51-f347063a55b5" Nov 24 14:34:59 crc kubenswrapper[4822]: I1124 14:34:59.038102 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" event={"ID":"ac525cf9-4bbf-4030-a7b3-f2da129d81c8","Type":"ContainerStarted","Data":"f9045f2ec32dd16bcc1806e4319698cfd09c4d06c9a5d75792b71d76a9f27e1c"} Nov 24 14:34:59 crc kubenswrapper[4822]: E1124 14:34:59.040879 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" podUID="ac525cf9-4bbf-4030-a7b3-f2da129d81c8" Nov 24 14:34:59 crc kubenswrapper[4822]: I1124 14:34:59.045167 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" event={"ID":"a648d53c-be72-46ed-8308-79565a88bd4c","Type":"ContainerStarted","Data":"47a0e80a550d391de8ede40684944e248132ea3c0c0ce4e72328eb656c18df84"} Nov 24 14:34:59 crc kubenswrapper[4822]: I1124 14:34:59.045273 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" event={"ID":"a648d53c-be72-46ed-8308-79565a88bd4c","Type":"ContainerStarted","Data":"ebce26cf8f3f3201772650f99bf73642392da408864a805f0c1016098fdc92ce"} Nov 24 14:34:59 crc kubenswrapper[4822]: I1124 14:34:59.045289 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" event={"ID":"a648d53c-be72-46ed-8308-79565a88bd4c","Type":"ContainerStarted","Data":"45b34f8767720ab2a3bdb3f98dd28bda7c3d18d09c94d9be24dfdc807257d018"} Nov 24 14:34:59 crc kubenswrapper[4822]: I1124 14:34:59.045472 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:34:59 crc kubenswrapper[4822]: E1124 14:34:59.046364 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" podUID="4e5bcbaf-c685-4047-a37a-3c058f8e1cbc" Nov 24 14:34:59 crc kubenswrapper[4822]: I1124 14:34:59.145450 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" podStartSLOduration=4.145425048 podStartE2EDuration="4.145425048s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:34:59.116304654 +0000 UTC m=+936.232945141" watchObservedRunningTime="2025-11-24 14:34:59.145425048 +0000 UTC m=+936.262065525" Nov 24 14:35:00 crc kubenswrapper[4822]: E1124 14:35:00.061132 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" podUID="3e001a57-d6f1-4d8a-9d51-f347063a55b5" Nov 24 14:35:00 crc kubenswrapper[4822]: E1124 14:35:00.061223 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" podUID="ac525cf9-4bbf-4030-a7b3-f2da129d81c8" Nov 24 14:35:00 crc kubenswrapper[4822]: E1124 14:35:00.061597 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" podUID="980ad0a7-70cd-4f77-bc29-119868034087" Nov 24 14:35:00 crc kubenswrapper[4822]: E1124 14:35:00.061714 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" podUID="97913a25-5fcf-4541-b928-191ffe821e93" Nov 24 14:35:00 crc kubenswrapper[4822]: E1124 14:35:00.061763 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" podUID="a22d61a1-a9c4-47e9-8c5d-6593642e798b" Nov 24 14:35:00 crc kubenswrapper[4822]: E1124 14:35:00.061825 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.73:5001/openstack-k8s-operators/telemetry-operator:57284e36a25ee63dd9ec7bcaee4ba1f50bec5c34\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" podUID="253630d1-2b9c-4bc4-a7fd-357c66d00f9e" Nov 24 14:35:07 crc kubenswrapper[4822]: I1124 14:35:07.167169 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7b58687767-tfthk" Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.171907 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" event={"ID":"e260fa6b-7576-4e48-8025-2c7ef2bd97f7","Type":"ContainerStarted","Data":"bbdb96c7b1874a8aec6f8ea1f83f5850f3a5134233f22f9e36ad9dc462c569ed"} Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.178878 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" event={"ID":"38159b16-c3af-4bcd-8734-88cfef861663","Type":"ContainerStarted","Data":"1da3f76a1fd94a9f74e177328a94815b16d1b4fc6e0b864c2bd79a94299c7a9d"} Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.180293 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" event={"ID":"10f3ea0d-a23e-47eb-b344-0696f8c9b1e8","Type":"ContainerStarted","Data":"3baab02f44f52668f02bdcd3bb548e18aa74e0cd7f8934cbe713a809e171105d"} Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.184978 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" event={"ID":"3809d742-f676-41dd-aae2-d03be501e0b0","Type":"ContainerStarted","Data":"57da1063b7e715f7aaaefe71e8bb7bfad7db96179d72eb4bb4928013e60bcd4f"} Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.194364 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" event={"ID":"b5737c88-eb91-432d-a620-734337051260","Type":"ContainerStarted","Data":"f6619f26fbb2c84714eb9032c26e76c37f6215beb522e09677a2459768d2f1b1"} Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.201528 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" event={"ID":"fd0cee72-09cc-449b-a6f7-fb951f0b87ba","Type":"ContainerStarted","Data":"136e36f17e0fb2db4412a4f6d52215df038e45deada2ed093a86ef6e34c4e75f"} Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.207624 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" event={"ID":"37a54972-a29e-4ae1-a3f1-104bc634b3cc","Type":"ContainerStarted","Data":"b98e7e728867590a145464cc708241004c0a26ca911c5af5d6c530e92d730484"} Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.224939 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" event={"ID":"dbffd719-e3e2-4728-9301-da9efe8eb57b","Type":"ContainerStarted","Data":"e3992b982db8439813a2c417065d98827b747ba7a2968345df8611a12d007de0"} Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.229974 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" event={"ID":"57aa63d3-2464-4cc7-a5e1-539e60ee6cca","Type":"ContainerStarted","Data":"e406a546b51eccc8955956369543ad58b05eef5a9262107504292836867f88d9"} Nov 24 14:35:08 crc kubenswrapper[4822]: I1124 14:35:08.243917 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" event={"ID":"b01f0f96-a40f-407a-a610-bd2e945a3e49","Type":"ContainerStarted","Data":"d609e79cff425a8d581319a11bd2e485f3b4e599d41faa44e1544aabfd4a9a30"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.262915 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" event={"ID":"3809d742-f676-41dd-aae2-d03be501e0b0","Type":"ContainerStarted","Data":"0873d9326d67f94475c2c81fc33f280f98561626df1a4353f6c5d0c6c2262f38"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.263436 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.276100 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" event={"ID":"e260fa6b-7576-4e48-8025-2c7ef2bd97f7","Type":"ContainerStarted","Data":"9d21142df24427c347abb616ec2f03b70545037fb85611924f8448f2cd9691a8"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.276985 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.284597 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" event={"ID":"37a54972-a29e-4ae1-a3f1-104bc634b3cc","Type":"ContainerStarted","Data":"db021476f4fc497d268346542fa9faaa2c8f908daa66e5d11c34caadbc5d042d"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.284838 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.293941 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" event={"ID":"b5737c88-eb91-432d-a620-734337051260","Type":"ContainerStarted","Data":"741c82a1257e13e1b20b12da670ed779d66d4aa04624968330a33559732127fb"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.297497 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.315488 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" event={"ID":"10f3ea0d-a23e-47eb-b344-0696f8c9b1e8","Type":"ContainerStarted","Data":"786788930d7afe99e93fba891f418603aeab39682197cd9b569b6fe3d7231747"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.316364 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.327903 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" event={"ID":"5047b32c-bd21-4f83-bce1-248a1ce109e9","Type":"ContainerStarted","Data":"11bfbcd31474f65673718a809c09b9e775638c216e367ad0278c174548a609b1"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.327932 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" event={"ID":"5047b32c-bd21-4f83-bce1-248a1ce109e9","Type":"ContainerStarted","Data":"4404ed544bd4cd8334d3311a855c42d6de878bdf63adab78339b093869130f9c"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.328529 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.335693 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" podStartSLOduration=4.214872951 podStartE2EDuration="14.3356645s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.158867349 +0000 UTC m=+934.275507826" lastFinishedPulling="2025-11-24 14:35:07.279658898 +0000 UTC m=+944.396299375" observedRunningTime="2025-11-24 14:35:09.305845915 +0000 UTC m=+946.422486402" watchObservedRunningTime="2025-11-24 14:35:09.3356645 +0000 UTC m=+946.452304977" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.339726 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" podStartSLOduration=4.582166476 podStartE2EDuration="14.339715749s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.573625309 +0000 UTC m=+934.690265776" lastFinishedPulling="2025-11-24 14:35:07.331174572 +0000 UTC m=+944.447815049" observedRunningTime="2025-11-24 14:35:09.328559565 +0000 UTC m=+946.445200052" watchObservedRunningTime="2025-11-24 14:35:09.339715749 +0000 UTC m=+946.456356226" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.342477 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" event={"ID":"38159b16-c3af-4bcd-8734-88cfef861663","Type":"ContainerStarted","Data":"1a66eb8ee399e8ab73fa282a2745c781c2683077ed0738d159a459c773b4aa01"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.343019 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.349145 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" event={"ID":"80740002-5d02-4aa7-b6fa-3de642266f21","Type":"ContainerStarted","Data":"3fcd04db59fd9a5accae7213bbdd5d00025d26c9f827ab6e0d430a8a01be1802"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.349189 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" event={"ID":"80740002-5d02-4aa7-b6fa-3de642266f21","Type":"ContainerStarted","Data":"4ec1d812c9f584afd960521cd4507f0b24c060fd9691f1aba5f0220d0d23a448"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.349850 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.358738 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" event={"ID":"8939f780-02b5-4022-80cb-a6fd247aa7f8","Type":"ContainerStarted","Data":"bf7fe44dcec3133416290ceab4e3c3b8a83dd905e637945f53eb4a6ee5fa8eff"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.358772 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" event={"ID":"8939f780-02b5-4022-80cb-a6fd247aa7f8","Type":"ContainerStarted","Data":"bebc0e5a77ad96239f8fc3b1990ac35747240034623f75f4f8d9dc87114c3529"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.359195 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.365361 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" podStartSLOduration=3.739804977 podStartE2EDuration="14.365337161s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:56.634915445 +0000 UTC m=+933.751555912" lastFinishedPulling="2025-11-24 14:35:07.260447619 +0000 UTC m=+944.377088096" observedRunningTime="2025-11-24 14:35:09.358951739 +0000 UTC m=+946.475592226" watchObservedRunningTime="2025-11-24 14:35:09.365337161 +0000 UTC m=+946.481977638" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.369256 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" event={"ID":"d758b1f7-956b-4881-8a61-a81e55fdffcd","Type":"ContainerStarted","Data":"0c378ad90aa16e6636ea1762592fd47af21d6997ecc89a4f5dbb99683824e1a8"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.369297 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" event={"ID":"d758b1f7-956b-4881-8a61-a81e55fdffcd","Type":"ContainerStarted","Data":"42301cc52eba0ddc00e9753219247a8915cdfd182b10253a7d7af6eeee5c343a"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.369958 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.371275 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" event={"ID":"dbffd719-e3e2-4728-9301-da9efe8eb57b","Type":"ContainerStarted","Data":"175cca5ae07011d05961356d9cbca1a475ca7edd22c313e7f46f0f10219f32df"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.371490 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.376883 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" event={"ID":"70ff7dde-abd4-453c-8164-c7304292c41d","Type":"ContainerStarted","Data":"e2a85c3137a1ef7d1c6c0f3646bb7b56a55092007ef1d77708d0148d64616716"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.376910 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" event={"ID":"70ff7dde-abd4-453c-8164-c7304292c41d","Type":"ContainerStarted","Data":"aeae3b740313ad5f14530878de3a81a38799f7c6357e01e7fcbc0f48e32f92e9"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.377012 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.385302 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" event={"ID":"57aa63d3-2464-4cc7-a5e1-539e60ee6cca","Type":"ContainerStarted","Data":"02f5edc96f16bfd36678c17d5c50a48c314833f0c25dd6ebd380091aa736a170"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.386014 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.396381 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" event={"ID":"b01f0f96-a40f-407a-a610-bd2e945a3e49","Type":"ContainerStarted","Data":"a8f8010cdcb3215b78011e6f6b431619a246b665ef344add1e647979a9abe7c9"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.397082 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.399540 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" event={"ID":"fd0cee72-09cc-449b-a6f7-fb951f0b87ba","Type":"ContainerStarted","Data":"b4dc4e3ff51f4cd1ae7dc8f72468367c89a6695f80a7ce0be14496f857f5ef04"} Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.399931 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.410012 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" podStartSLOduration=4.674977019 podStartE2EDuration="14.409985777s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.515978131 +0000 UTC m=+934.632618608" lastFinishedPulling="2025-11-24 14:35:07.250986869 +0000 UTC m=+944.367627366" observedRunningTime="2025-11-24 14:35:09.409409189 +0000 UTC m=+946.526049676" watchObservedRunningTime="2025-11-24 14:35:09.409985777 +0000 UTC m=+946.526626254" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.413581 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" podStartSLOduration=4.679314946 podStartE2EDuration="14.41357446s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.516656043 +0000 UTC m=+934.633296520" lastFinishedPulling="2025-11-24 14:35:07.250915557 +0000 UTC m=+944.367556034" observedRunningTime="2025-11-24 14:35:09.381095881 +0000 UTC m=+946.497736368" watchObservedRunningTime="2025-11-24 14:35:09.41357446 +0000 UTC m=+946.530214937" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.452722 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" podStartSLOduration=4.114159047 podStartE2EDuration="14.452699121s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:56.991501531 +0000 UTC m=+934.108142008" lastFinishedPulling="2025-11-24 14:35:07.330041605 +0000 UTC m=+944.446682082" observedRunningTime="2025-11-24 14:35:09.450600375 +0000 UTC m=+946.567240892" watchObservedRunningTime="2025-11-24 14:35:09.452699121 +0000 UTC m=+946.569339598" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.482188 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" podStartSLOduration=4.699754615 podStartE2EDuration="14.482165416s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.574535538 +0000 UTC m=+934.691176015" lastFinishedPulling="2025-11-24 14:35:07.356946329 +0000 UTC m=+944.473586816" observedRunningTime="2025-11-24 14:35:09.477020133 +0000 UTC m=+946.593660640" watchObservedRunningTime="2025-11-24 14:35:09.482165416 +0000 UTC m=+946.598805893" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.529299 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" podStartSLOduration=4.785819905 podStartE2EDuration="14.529272849s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.517041026 +0000 UTC m=+934.633681503" lastFinishedPulling="2025-11-24 14:35:07.26049394 +0000 UTC m=+944.377134447" observedRunningTime="2025-11-24 14:35:09.507755467 +0000 UTC m=+946.624395954" watchObservedRunningTime="2025-11-24 14:35:09.529272849 +0000 UTC m=+946.645913326" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.530856 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" podStartSLOduration=4.2691570500000005 podStartE2EDuration="14.530847379s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:56.997872973 +0000 UTC m=+934.114513450" lastFinishedPulling="2025-11-24 14:35:07.259563282 +0000 UTC m=+944.376203779" observedRunningTime="2025-11-24 14:35:09.526884694 +0000 UTC m=+946.643525181" watchObservedRunningTime="2025-11-24 14:35:09.530847379 +0000 UTC m=+946.647487856" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.556064 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" podStartSLOduration=4.814634817 podStartE2EDuration="14.556039447s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.57553301 +0000 UTC m=+934.692173487" lastFinishedPulling="2025-11-24 14:35:07.31693763 +0000 UTC m=+944.433578117" observedRunningTime="2025-11-24 14:35:09.548470238 +0000 UTC m=+946.665110725" watchObservedRunningTime="2025-11-24 14:35:09.556039447 +0000 UTC m=+946.672679934" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.575256 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" podStartSLOduration=3.920570627 podStartE2EDuration="14.575232056s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:56.666505706 +0000 UTC m=+933.783146183" lastFinishedPulling="2025-11-24 14:35:07.321167135 +0000 UTC m=+944.437807612" observedRunningTime="2025-11-24 14:35:09.567360496 +0000 UTC m=+946.684000983" watchObservedRunningTime="2025-11-24 14:35:09.575232056 +0000 UTC m=+946.691872533" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.586968 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" podStartSLOduration=4.844805113 podStartE2EDuration="14.586944487s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.577692228 +0000 UTC m=+934.694332705" lastFinishedPulling="2025-11-24 14:35:07.319831592 +0000 UTC m=+944.436472079" observedRunningTime="2025-11-24 14:35:09.584159919 +0000 UTC m=+946.700800406" watchObservedRunningTime="2025-11-24 14:35:09.586944487 +0000 UTC m=+946.703584964" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.607717 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" podStartSLOduration=4.871123489 podStartE2EDuration="14.607694166s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.578199225 +0000 UTC m=+934.694839702" lastFinishedPulling="2025-11-24 14:35:07.314769902 +0000 UTC m=+944.431410379" observedRunningTime="2025-11-24 14:35:09.603242084 +0000 UTC m=+946.719882581" watchObservedRunningTime="2025-11-24 14:35:09.607694166 +0000 UTC m=+946.724334643" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.633372 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" podStartSLOduration=4.450150511 podStartE2EDuration="14.633346979s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.00123203 +0000 UTC m=+934.117872507" lastFinishedPulling="2025-11-24 14:35:07.184428458 +0000 UTC m=+944.301068975" observedRunningTime="2025-11-24 14:35:09.626700358 +0000 UTC m=+946.743340835" watchObservedRunningTime="2025-11-24 14:35:09.633346979 +0000 UTC m=+946.749987456" Nov 24 14:35:09 crc kubenswrapper[4822]: I1124 14:35:09.662334 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" podStartSLOduration=5.221376864 podStartE2EDuration="14.662313287s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.78240844 +0000 UTC m=+934.899048917" lastFinishedPulling="2025-11-24 14:35:07.223344843 +0000 UTC m=+944.339985340" observedRunningTime="2025-11-24 14:35:09.657258277 +0000 UTC m=+946.773898774" watchObservedRunningTime="2025-11-24 14:35:09.662313287 +0000 UTC m=+946.778953764" Nov 24 14:35:11 crc kubenswrapper[4822]: I1124 14:35:11.577973 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:35:11 crc kubenswrapper[4822]: I1124 14:35:11.578479 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:35:11 crc kubenswrapper[4822]: I1124 14:35:11.578556 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:35:11 crc kubenswrapper[4822]: I1124 14:35:11.579665 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb8615da8ae4badf56e6f90df575844458ac64f06801fa8ab7085476265bc5df"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:35:11 crc kubenswrapper[4822]: I1124 14:35:11.579775 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://cb8615da8ae4badf56e6f90df575844458ac64f06801fa8ab7085476265bc5df" gracePeriod=600 Nov 24 14:35:13 crc kubenswrapper[4822]: I1124 14:35:13.460675 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="cb8615da8ae4badf56e6f90df575844458ac64f06801fa8ab7085476265bc5df" exitCode=0 Nov 24 14:35:13 crc kubenswrapper[4822]: I1124 14:35:13.460762 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"cb8615da8ae4badf56e6f90df575844458ac64f06801fa8ab7085476265bc5df"} Nov 24 14:35:13 crc kubenswrapper[4822]: I1124 14:35:13.460837 4822 scope.go:117] "RemoveContainer" containerID="065577b1f1af0833feb9bb8a7aed6dd3738bba738b7c830369d5c019d2095f70" Nov 24 14:35:15 crc kubenswrapper[4822]: I1124 14:35:15.428635 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-pvs2h" Nov 24 14:35:15 crc kubenswrapper[4822]: I1124 14:35:15.443037 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-6f4n4" Nov 24 14:35:15 crc kubenswrapper[4822]: I1124 14:35:15.455001 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-tvqv6" Nov 24 14:35:15 crc kubenswrapper[4822]: I1124 14:35:15.496076 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xh9nm" Nov 24 14:35:15 crc kubenswrapper[4822]: I1124 14:35:15.523859 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4fgsn" Nov 24 14:35:15 crc kubenswrapper[4822]: I1124 14:35:15.544772 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-k6hqr" Nov 24 14:35:15 crc kubenswrapper[4822]: I1124 14:35:15.923437 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-khnxm" Nov 24 14:35:15 crc kubenswrapper[4822]: I1124 14:35:15.950686 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-bzj92" Nov 24 14:35:16 crc kubenswrapper[4822]: I1124 14:35:16.021190 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-xd82t" Nov 24 14:35:16 crc kubenswrapper[4822]: I1124 14:35:16.089692 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-mf4hq" Nov 24 14:35:16 crc kubenswrapper[4822]: I1124 14:35:16.109889 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-lsbgj" Nov 24 14:35:16 crc kubenswrapper[4822]: I1124 14:35:16.237941 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-45wdc" Nov 24 14:35:16 crc kubenswrapper[4822]: I1124 14:35:16.261841 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-9j9ls" Nov 24 14:35:16 crc kubenswrapper[4822]: I1124 14:35:16.291790 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-wtv4t" Nov 24 14:35:16 crc kubenswrapper[4822]: I1124 14:35:16.528654 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"eb98f980c90c3e30ffa6cd4e582af8567a0ef6fab6806bdeaf156e64abf982bb"} Nov 24 14:35:16 crc kubenswrapper[4822]: I1124 14:35:16.684544 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.616663 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" event={"ID":"253630d1-2b9c-4bc4-a7fd-357c66d00f9e","Type":"ContainerStarted","Data":"2662feb7d5e61d1e4a56cc683ba01280446fa23ac561ee28104bba7b813d66b4"} Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.617964 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.618956 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" event={"ID":"980ad0a7-70cd-4f77-bc29-119868034087","Type":"ContainerStarted","Data":"4f7c8e5b8e3dcaca9042ac43a1edf3a20399789466660d2ab00e2f3d8625d64b"} Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.619199 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.622126 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" event={"ID":"a22d61a1-a9c4-47e9-8c5d-6593642e798b","Type":"ContainerStarted","Data":"bca777871c861f8e37a6630be7a98db55e995be90a4ff20afcf6d4a348ac2f81"} Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.622578 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.624046 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" event={"ID":"3e001a57-d6f1-4d8a-9d51-f347063a55b5","Type":"ContainerStarted","Data":"190ba9d92866f244a394d8c4fc3e11614395f89e230cc82e898eea66153b24da"} Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.624255 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.628349 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" event={"ID":"4e5bcbaf-c685-4047-a37a-3c058f8e1cbc","Type":"ContainerStarted","Data":"efeb2f60faac289ac9e0a151a6729fbc1ec8a80a946734d5b1b4035845c44beb"} Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.633535 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" event={"ID":"ac525cf9-4bbf-4030-a7b3-f2da129d81c8","Type":"ContainerStarted","Data":"6a83a46c04f44624a4562e06fce4cbfc69682f0e52c8e08e0e9745efb2df2791"} Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.633960 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.638559 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" podStartSLOduration=3.755226107 podStartE2EDuration="29.638536854s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.633908821 +0000 UTC m=+934.750549298" lastFinishedPulling="2025-11-24 14:35:23.517219568 +0000 UTC m=+960.633860045" observedRunningTime="2025-11-24 14:35:24.635414665 +0000 UTC m=+961.752055142" watchObservedRunningTime="2025-11-24 14:35:24.638536854 +0000 UTC m=+961.755177331" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.646476 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" event={"ID":"97913a25-5fcf-4541-b928-191ffe821e93","Type":"ContainerStarted","Data":"6fca1aede7bb2ad77a96519a11bb54fe2ab4c854c203a82534948970f66985ee"} Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.647706 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.677595 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" podStartSLOduration=3.735022065 podStartE2EDuration="29.677543751s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.578259866 +0000 UTC m=+934.694900343" lastFinishedPulling="2025-11-24 14:35:23.520781552 +0000 UTC m=+960.637422029" observedRunningTime="2025-11-24 14:35:24.662774032 +0000 UTC m=+961.779414529" watchObservedRunningTime="2025-11-24 14:35:24.677543751 +0000 UTC m=+961.794184228" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.696417 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" podStartSLOduration=3.767879166 podStartE2EDuration="29.696395908s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.58847697 +0000 UTC m=+934.705117447" lastFinishedPulling="2025-11-24 14:35:23.516993672 +0000 UTC m=+960.633634189" observedRunningTime="2025-11-24 14:35:24.690008976 +0000 UTC m=+961.806649483" watchObservedRunningTime="2025-11-24 14:35:24.696395908 +0000 UTC m=+961.813036395" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.711075 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" podStartSLOduration=3.789359548 podStartE2EDuration="29.711059033s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.595171863 +0000 UTC m=+934.711812340" lastFinishedPulling="2025-11-24 14:35:23.516871348 +0000 UTC m=+960.633511825" observedRunningTime="2025-11-24 14:35:24.707434368 +0000 UTC m=+961.824074855" watchObservedRunningTime="2025-11-24 14:35:24.711059033 +0000 UTC m=+961.827699510" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.732833 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" podStartSLOduration=3.797857167 podStartE2EDuration="29.732809213s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.583591015 +0000 UTC m=+934.700231492" lastFinishedPulling="2025-11-24 14:35:23.518543061 +0000 UTC m=+960.635183538" observedRunningTime="2025-11-24 14:35:24.731085238 +0000 UTC m=+961.847725735" watchObservedRunningTime="2025-11-24 14:35:24.732809213 +0000 UTC m=+961.849449690" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.751995 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-98xcn" podStartSLOduration=3.920417393 podStartE2EDuration="29.75197187s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.820040852 +0000 UTC m=+934.936681319" lastFinishedPulling="2025-11-24 14:35:23.651595319 +0000 UTC m=+960.768235796" observedRunningTime="2025-11-24 14:35:24.746623331 +0000 UTC m=+961.863263818" watchObservedRunningTime="2025-11-24 14:35:24.75197187 +0000 UTC m=+961.868612347" Nov 24 14:35:24 crc kubenswrapper[4822]: I1124 14:35:24.770080 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" podStartSLOduration=3.866835095 podStartE2EDuration="29.770058344s" podCreationTimestamp="2025-11-24 14:34:55 +0000 UTC" firstStartedPulling="2025-11-24 14:34:57.614509486 +0000 UTC m=+934.731149963" lastFinishedPulling="2025-11-24 14:35:23.517732735 +0000 UTC m=+960.634373212" observedRunningTime="2025-11-24 14:35:24.762615258 +0000 UTC m=+961.879255755" watchObservedRunningTime="2025-11-24 14:35:24.770058344 +0000 UTC m=+961.886698821" Nov 24 14:35:35 crc kubenswrapper[4822]: I1124 14:35:35.796823 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-z8mj6" Nov 24 14:35:35 crc kubenswrapper[4822]: I1124 14:35:35.855112 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-2vr5g" Nov 24 14:35:35 crc kubenswrapper[4822]: I1124 14:35:35.936269 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-jkq8m" Nov 24 14:35:36 crc kubenswrapper[4822]: I1124 14:35:36.365082 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-54d7678447-dx47k" Nov 24 14:35:36 crc kubenswrapper[4822]: I1124 14:35:36.512184 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-2p2qp" Nov 24 14:35:36 crc kubenswrapper[4822]: I1124 14:35:36.529729 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-dc279" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.668079 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4ktn"] Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.672794 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.675826 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.676120 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-vxx5f" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.676257 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.683558 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.684650 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4ktn"] Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.718928 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9mnsx"] Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.720309 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.726587 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.730996 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9mnsx"] Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.862079 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55hrz\" (UniqueName: \"kubernetes.io/projected/76db3a36-9f32-48e9-930d-69aee5393aa4-kube-api-access-55hrz\") pod \"dnsmasq-dns-78dd6ddcc-9mnsx\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.862174 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-config\") pod \"dnsmasq-dns-78dd6ddcc-9mnsx\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.862200 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwv8f\" (UniqueName: \"kubernetes.io/projected/552d1d02-3200-45d0-a913-45c5188163b9-kube-api-access-cwv8f\") pod \"dnsmasq-dns-675f4bcbfc-q4ktn\" (UID: \"552d1d02-3200-45d0-a913-45c5188163b9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.862240 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9mnsx\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.862326 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552d1d02-3200-45d0-a913-45c5188163b9-config\") pod \"dnsmasq-dns-675f4bcbfc-q4ktn\" (UID: \"552d1d02-3200-45d0-a913-45c5188163b9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.963221 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-config\") pod \"dnsmasq-dns-78dd6ddcc-9mnsx\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.963376 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwv8f\" (UniqueName: \"kubernetes.io/projected/552d1d02-3200-45d0-a913-45c5188163b9-kube-api-access-cwv8f\") pod \"dnsmasq-dns-675f4bcbfc-q4ktn\" (UID: \"552d1d02-3200-45d0-a913-45c5188163b9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.963619 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9mnsx\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.963748 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552d1d02-3200-45d0-a913-45c5188163b9-config\") pod \"dnsmasq-dns-675f4bcbfc-q4ktn\" (UID: \"552d1d02-3200-45d0-a913-45c5188163b9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.963807 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55hrz\" (UniqueName: \"kubernetes.io/projected/76db3a36-9f32-48e9-930d-69aee5393aa4-kube-api-access-55hrz\") pod \"dnsmasq-dns-78dd6ddcc-9mnsx\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.964548 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9mnsx\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.964815 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552d1d02-3200-45d0-a913-45c5188163b9-config\") pod \"dnsmasq-dns-675f4bcbfc-q4ktn\" (UID: \"552d1d02-3200-45d0-a913-45c5188163b9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.964824 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-config\") pod \"dnsmasq-dns-78dd6ddcc-9mnsx\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.985106 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwv8f\" (UniqueName: \"kubernetes.io/projected/552d1d02-3200-45d0-a913-45c5188163b9-kube-api-access-cwv8f\") pod \"dnsmasq-dns-675f4bcbfc-q4ktn\" (UID: \"552d1d02-3200-45d0-a913-45c5188163b9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:35:55 crc kubenswrapper[4822]: I1124 14:35:55.987785 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55hrz\" (UniqueName: \"kubernetes.io/projected/76db3a36-9f32-48e9-930d-69aee5393aa4-kube-api-access-55hrz\") pod \"dnsmasq-dns-78dd6ddcc-9mnsx\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:56 crc kubenswrapper[4822]: I1124 14:35:56.006787 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:35:56 crc kubenswrapper[4822]: I1124 14:35:56.047190 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:35:56 crc kubenswrapper[4822]: I1124 14:35:56.504239 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4ktn"] Nov 24 14:35:56 crc kubenswrapper[4822]: I1124 14:35:56.576721 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9mnsx"] Nov 24 14:35:56 crc kubenswrapper[4822]: W1124 14:35:56.579966 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76db3a36_9f32_48e9_930d_69aee5393aa4.slice/crio-9f1a4969cc0ea25440ae085977a30791c7cc16a96eff13b20b349206a86d258a WatchSource:0}: Error finding container 9f1a4969cc0ea25440ae085977a30791c7cc16a96eff13b20b349206a86d258a: Status 404 returned error can't find the container with id 9f1a4969cc0ea25440ae085977a30791c7cc16a96eff13b20b349206a86d258a Nov 24 14:35:56 crc kubenswrapper[4822]: I1124 14:35:56.979470 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" event={"ID":"552d1d02-3200-45d0-a913-45c5188163b9","Type":"ContainerStarted","Data":"3d5362d26e793b20896b661855be65fb83c50336705cb8f8bca4503932a7c7cb"} Nov 24 14:35:56 crc kubenswrapper[4822]: I1124 14:35:56.980972 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" event={"ID":"76db3a36-9f32-48e9-930d-69aee5393aa4","Type":"ContainerStarted","Data":"9f1a4969cc0ea25440ae085977a30791c7cc16a96eff13b20b349206a86d258a"} Nov 24 14:35:58 crc kubenswrapper[4822]: I1124 14:35:58.931522 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4ktn"] Nov 24 14:35:58 crc kubenswrapper[4822]: I1124 14:35:58.951894 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vd528"] Nov 24 14:35:58 crc kubenswrapper[4822]: I1124 14:35:58.957848 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.003143 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vd528"] Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.008974 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vd528\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.009121 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ggpf\" (UniqueName: \"kubernetes.io/projected/8f960c65-fc33-4f09-a525-e0542ce7207c-kube-api-access-2ggpf\") pod \"dnsmasq-dns-666b6646f7-vd528\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.009369 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-config\") pod \"dnsmasq-dns-666b6646f7-vd528\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.110725 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ggpf\" (UniqueName: \"kubernetes.io/projected/8f960c65-fc33-4f09-a525-e0542ce7207c-kube-api-access-2ggpf\") pod \"dnsmasq-dns-666b6646f7-vd528\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.110778 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-config\") pod \"dnsmasq-dns-666b6646f7-vd528\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.110869 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vd528\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.111719 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vd528\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.112010 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-config\") pod \"dnsmasq-dns-666b6646f7-vd528\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.136545 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ggpf\" (UniqueName: \"kubernetes.io/projected/8f960c65-fc33-4f09-a525-e0542ce7207c-kube-api-access-2ggpf\") pod \"dnsmasq-dns-666b6646f7-vd528\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.218613 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9mnsx"] Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.238560 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfn5c"] Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.242699 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.251972 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfn5c"] Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.303285 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.414745 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cfn5c\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.414803 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nj6g\" (UniqueName: \"kubernetes.io/projected/a16b9b73-6577-4420-89bb-3c774f9c4bd3-kube-api-access-2nj6g\") pod \"dnsmasq-dns-57d769cc4f-cfn5c\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.414885 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-config\") pod \"dnsmasq-dns-57d769cc4f-cfn5c\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.517509 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cfn5c\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.517562 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nj6g\" (UniqueName: \"kubernetes.io/projected/a16b9b73-6577-4420-89bb-3c774f9c4bd3-kube-api-access-2nj6g\") pod \"dnsmasq-dns-57d769cc4f-cfn5c\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.517656 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-config\") pod \"dnsmasq-dns-57d769cc4f-cfn5c\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.523102 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cfn5c\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.523177 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-config\") pod \"dnsmasq-dns-57d769cc4f-cfn5c\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.535569 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nj6g\" (UniqueName: \"kubernetes.io/projected/a16b9b73-6577-4420-89bb-3c774f9c4bd3-kube-api-access-2nj6g\") pod \"dnsmasq-dns-57d769cc4f-cfn5c\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:35:59 crc kubenswrapper[4822]: I1124 14:35:59.592503 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.103384 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.106550 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.109117 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.109466 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.110261 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.113449 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.114686 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.114809 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.114955 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.115034 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vc8w9" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228329 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228399 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228428 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlljz\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-kube-api-access-nlljz\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228483 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228681 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228749 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228789 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228849 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-config-data\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228871 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228942 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.228998 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.330993 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331049 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331078 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331099 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlljz\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-kube-api-access-nlljz\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331135 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331170 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331191 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331222 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331246 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-config-data\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331263 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331288 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.331692 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.332509 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.333476 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.334088 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-config-data\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.334865 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.336045 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.336880 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.336917 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0ec351afa2b1df6aac252f23371d09947c6973707300212e509127eecca4cbe8/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.352421 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.352431 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.353946 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.357678 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlljz\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-kube-api-access-nlljz\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.375337 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.376780 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.389694 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.395495 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") pod \"rabbitmq-server-0\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.430729 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.431106 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.431240 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.431738 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.431876 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.431982 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.436108 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-bzdzc" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.536814 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537101 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537123 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537145 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537166 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw4w4\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-kube-api-access-qw4w4\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537195 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537228 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a52d37ed-a00f-4983-88ee-023bd282038e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537249 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537271 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537291 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.537309 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a52d37ed-a00f-4983-88ee-023bd282038e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.640904 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a52d37ed-a00f-4983-88ee-023bd282038e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.640949 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.640978 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.641002 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.641023 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a52d37ed-a00f-4983-88ee-023bd282038e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.641151 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.641607 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.641665 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.641686 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.642598 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.641703 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.642706 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.642753 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw4w4\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-kube-api-access-qw4w4\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.642947 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.643338 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.644384 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.645134 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a52d37ed-a00f-4983-88ee-023bd282038e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.645687 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a52d37ed-a00f-4983-88ee-023bd282038e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.646303 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.647067 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.647096 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c6993c052f1ba2e354c4a7c732feee88a39bd89c27ed6adb11f5ec90b4236450/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.659966 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.675673 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw4w4\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-kube-api-access-qw4w4\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.697117 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") pod \"rabbitmq-cell1-server-0\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.724525 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:36:00 crc kubenswrapper[4822]: I1124 14:36:00.777304 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.786870 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.788103 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.792381 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.792520 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.792613 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-m4tm2" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.798922 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.801401 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.813468 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.963287 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a756bbea-7fc2-443e-8d92-37f993ad9ad8-kolla-config\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.963375 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a756bbea-7fc2-443e-8d92-37f993ad9ad8-config-data-default\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.963422 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3e643ef4-ef1e-4a8d-bab5-0408a3cef7be\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3e643ef4-ef1e-4a8d-bab5-0408a3cef7be\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.963463 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a756bbea-7fc2-443e-8d92-37f993ad9ad8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.963488 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a756bbea-7fc2-443e-8d92-37f993ad9ad8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.963513 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a756bbea-7fc2-443e-8d92-37f993ad9ad8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.963573 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br45z\" (UniqueName: \"kubernetes.io/projected/a756bbea-7fc2-443e-8d92-37f993ad9ad8-kube-api-access-br45z\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:01 crc kubenswrapper[4822]: I1124 14:36:01.963599 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a756bbea-7fc2-443e-8d92-37f993ad9ad8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.065064 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a756bbea-7fc2-443e-8d92-37f993ad9ad8-kolla-config\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.065148 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a756bbea-7fc2-443e-8d92-37f993ad9ad8-config-data-default\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.065196 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3e643ef4-ef1e-4a8d-bab5-0408a3cef7be\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3e643ef4-ef1e-4a8d-bab5-0408a3cef7be\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.065257 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a756bbea-7fc2-443e-8d92-37f993ad9ad8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.065283 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a756bbea-7fc2-443e-8d92-37f993ad9ad8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.065306 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a756bbea-7fc2-443e-8d92-37f993ad9ad8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.065365 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br45z\" (UniqueName: \"kubernetes.io/projected/a756bbea-7fc2-443e-8d92-37f993ad9ad8-kube-api-access-br45z\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.065390 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a756bbea-7fc2-443e-8d92-37f993ad9ad8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.066715 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a756bbea-7fc2-443e-8d92-37f993ad9ad8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.067327 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a756bbea-7fc2-443e-8d92-37f993ad9ad8-config-data-default\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.067834 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a756bbea-7fc2-443e-8d92-37f993ad9ad8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.068064 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a756bbea-7fc2-443e-8d92-37f993ad9ad8-kolla-config\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.070501 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a756bbea-7fc2-443e-8d92-37f993ad9ad8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.070880 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a756bbea-7fc2-443e-8d92-37f993ad9ad8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.072024 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.072062 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3e643ef4-ef1e-4a8d-bab5-0408a3cef7be\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3e643ef4-ef1e-4a8d-bab5-0408a3cef7be\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2cfd9fde1e18963d1dd39a306376c94f492e799f2bb016e667ad387e38b58f37/globalmount\"" pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.087905 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br45z\" (UniqueName: \"kubernetes.io/projected/a756bbea-7fc2-443e-8d92-37f993ad9ad8-kube-api-access-br45z\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.113642 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3e643ef4-ef1e-4a8d-bab5-0408a3cef7be\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3e643ef4-ef1e-4a8d-bab5-0408a3cef7be\") pod \"openstack-galera-0\" (UID: \"a756bbea-7fc2-443e-8d92-37f993ad9ad8\") " pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.406289 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.989653 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.991653 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.996604 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xpbv2" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.996784 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.996956 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 14:36:02 crc kubenswrapper[4822]: I1124 14:36:02.997059 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.003335 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.083972 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c617427-bd54-45c7-8e20-4ed74395e58c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.084080 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2sbt\" (UniqueName: \"kubernetes.io/projected/4c617427-bd54-45c7-8e20-4ed74395e58c-kube-api-access-l2sbt\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.084119 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c617427-bd54-45c7-8e20-4ed74395e58c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.084242 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c617427-bd54-45c7-8e20-4ed74395e58c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.084286 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-011ac132-2f3a-45f4-8ca0-08e5fb1c5673\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-011ac132-2f3a-45f4-8ca0-08e5fb1c5673\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.084316 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c617427-bd54-45c7-8e20-4ed74395e58c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.084394 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c617427-bd54-45c7-8e20-4ed74395e58c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.084447 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c617427-bd54-45c7-8e20-4ed74395e58c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.185760 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c617427-bd54-45c7-8e20-4ed74395e58c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.185858 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c617427-bd54-45c7-8e20-4ed74395e58c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.185916 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2sbt\" (UniqueName: \"kubernetes.io/projected/4c617427-bd54-45c7-8e20-4ed74395e58c-kube-api-access-l2sbt\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.185955 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c617427-bd54-45c7-8e20-4ed74395e58c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.186050 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c617427-bd54-45c7-8e20-4ed74395e58c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.186098 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-011ac132-2f3a-45f4-8ca0-08e5fb1c5673\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-011ac132-2f3a-45f4-8ca0-08e5fb1c5673\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.186133 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c617427-bd54-45c7-8e20-4ed74395e58c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.186233 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c617427-bd54-45c7-8e20-4ed74395e58c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.187606 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c617427-bd54-45c7-8e20-4ed74395e58c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.187823 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c617427-bd54-45c7-8e20-4ed74395e58c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.188796 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c617427-bd54-45c7-8e20-4ed74395e58c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.189063 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c617427-bd54-45c7-8e20-4ed74395e58c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.191354 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c617427-bd54-45c7-8e20-4ed74395e58c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.198445 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.198475 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-011ac132-2f3a-45f4-8ca0-08e5fb1c5673\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-011ac132-2f3a-45f4-8ca0-08e5fb1c5673\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4eb528ab7eef00b5d7d9f9b2bed8b3f00d654683e52699e1c452d95660fe623c/globalmount\"" pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.207307 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c617427-bd54-45c7-8e20-4ed74395e58c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.210787 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2sbt\" (UniqueName: \"kubernetes.io/projected/4c617427-bd54-45c7-8e20-4ed74395e58c-kube-api-access-l2sbt\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.245167 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-011ac132-2f3a-45f4-8ca0-08e5fb1c5673\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-011ac132-2f3a-45f4-8ca0-08e5fb1c5673\") pod \"openstack-cell1-galera-0\" (UID: \"4c617427-bd54-45c7-8e20-4ed74395e58c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.272925 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.273870 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.275856 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.276125 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.276297 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-ks8vj" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.287736 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.328414 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.388402 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/346023fa-f08b-4ddb-a527-a637be223953-kolla-config\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.388455 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jp7g\" (UniqueName: \"kubernetes.io/projected/346023fa-f08b-4ddb-a527-a637be223953-kube-api-access-5jp7g\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.388486 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/346023fa-f08b-4ddb-a527-a637be223953-combined-ca-bundle\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.388506 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/346023fa-f08b-4ddb-a527-a637be223953-memcached-tls-certs\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.388522 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/346023fa-f08b-4ddb-a527-a637be223953-config-data\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.489535 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/346023fa-f08b-4ddb-a527-a637be223953-kolla-config\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.489595 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jp7g\" (UniqueName: \"kubernetes.io/projected/346023fa-f08b-4ddb-a527-a637be223953-kube-api-access-5jp7g\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.489632 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/346023fa-f08b-4ddb-a527-a637be223953-combined-ca-bundle\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.489649 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/346023fa-f08b-4ddb-a527-a637be223953-memcached-tls-certs\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.489667 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/346023fa-f08b-4ddb-a527-a637be223953-config-data\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.490362 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/346023fa-f08b-4ddb-a527-a637be223953-kolla-config\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.490400 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/346023fa-f08b-4ddb-a527-a637be223953-config-data\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.493792 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/346023fa-f08b-4ddb-a527-a637be223953-combined-ca-bundle\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.496563 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/346023fa-f08b-4ddb-a527-a637be223953-memcached-tls-certs\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.505586 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jp7g\" (UniqueName: \"kubernetes.io/projected/346023fa-f08b-4ddb-a527-a637be223953-kube-api-access-5jp7g\") pod \"memcached-0\" (UID: \"346023fa-f08b-4ddb-a527-a637be223953\") " pod="openstack/memcached-0" Nov 24 14:36:03 crc kubenswrapper[4822]: I1124 14:36:03.591625 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.196719 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.197913 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.201518 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-gsl97" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.209760 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.326394 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7vzx\" (UniqueName: \"kubernetes.io/projected/e5326828-6e3f-464a-90d6-0cf579f5c598-kube-api-access-z7vzx\") pod \"kube-state-metrics-0\" (UID: \"e5326828-6e3f-464a-90d6-0cf579f5c598\") " pod="openstack/kube-state-metrics-0" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.427933 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7vzx\" (UniqueName: \"kubernetes.io/projected/e5326828-6e3f-464a-90d6-0cf579f5c598-kube-api-access-z7vzx\") pod \"kube-state-metrics-0\" (UID: \"e5326828-6e3f-464a-90d6-0cf579f5c598\") " pod="openstack/kube-state-metrics-0" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.476304 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7vzx\" (UniqueName: \"kubernetes.io/projected/e5326828-6e3f-464a-90d6-0cf579f5c598-kube-api-access-z7vzx\") pod \"kube-state-metrics-0\" (UID: \"e5326828-6e3f-464a-90d6-0cf579f5c598\") " pod="openstack/kube-state-metrics-0" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.519445 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.795081 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfn5c"] Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.937058 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.941020 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.942970 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.943245 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.943638 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-8jvdt" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.943671 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.943902 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Nov 24 14:36:05 crc kubenswrapper[4822]: I1124 14:36:05.952900 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.050170 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.050226 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khg9v\" (UniqueName: \"kubernetes.io/projected/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-kube-api-access-khg9v\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.050285 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.050304 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.050331 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.050370 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.050394 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.152237 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.152513 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.152559 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.152578 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khg9v\" (UniqueName: \"kubernetes.io/projected/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-kube-api-access-khg9v\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.152635 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.152651 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.152679 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.154131 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.158433 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.159146 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.159831 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.161637 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.164634 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.167987 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khg9v\" (UniqueName: \"kubernetes.io/projected/389f0d9e-a5ab-4d8b-82e2-f4c90d830c42-kube-api-access-khg9v\") pod \"alertmanager-metric-storage-0\" (UID: \"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.262770 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.510266 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.513543 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.520933 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.521242 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.521377 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-6pdph" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.521952 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.522299 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.522457 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.523701 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.558802 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.558869 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz778\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-kube-api-access-zz778\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.558919 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-config\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.558938 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/08f13b2a-d10d-4b51-be76-b4c510b24d00-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.559022 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.559055 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.559085 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.559131 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/08f13b2a-d10d-4b51-be76-b4c510b24d00-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.660333 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.660389 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.660422 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.660451 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/08f13b2a-d10d-4b51-be76-b4c510b24d00-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.660487 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.660527 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz778\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-kube-api-access-zz778\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.660553 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-config\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.660576 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/08f13b2a-d10d-4b51-be76-b4c510b24d00-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.661246 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/08f13b2a-d10d-4b51-be76-b4c510b24d00-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.664435 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/08f13b2a-d10d-4b51-be76-b4c510b24d00-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.667779 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-config\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.667856 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.668083 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.673796 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.673897 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.673920 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/af84fe80f7334969b1921a67001ef55fda3e7f45a332f15660d093709721d1df/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.679612 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz778\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-kube-api-access-zz778\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.706859 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") pod \"prometheus-metric-storage-0\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:06 crc kubenswrapper[4822]: I1124 14:36:06.848840 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.172323 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vh28h"] Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.173561 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.175504 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.176449 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.176598 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-zj52z" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.186843 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7l6s7"] Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.188530 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.204142 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vh28h"] Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.218446 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7l6s7"] Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.243479 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-ovn-controller-tls-certs\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.243553 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw6zk\" (UniqueName: \"kubernetes.io/projected/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-kube-api-access-bw6zk\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.243740 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-combined-ca-bundle\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.243798 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-var-run\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.244083 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-var-run-ovn\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.244154 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-scripts\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.244240 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-var-log-ovn\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346001 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-etc-ovs\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346063 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-var-log\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346100 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-var-run\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346169 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-var-run-ovn\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346195 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-scripts\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346240 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-var-log-ovn\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346270 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-ovn-controller-tls-certs\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346298 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-scripts\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346325 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw6zk\" (UniqueName: \"kubernetes.io/projected/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-kube-api-access-bw6zk\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346355 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trcgx\" (UniqueName: \"kubernetes.io/projected/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-kube-api-access-trcgx\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346373 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-var-lib\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346394 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-combined-ca-bundle\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346413 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-var-run\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346819 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-var-log-ovn\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.346939 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-var-run\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.347008 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-var-run-ovn\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.348265 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-scripts\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.352390 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-ovn-controller-tls-certs\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.353637 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-combined-ca-bundle\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.366101 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw6zk\" (UniqueName: \"kubernetes.io/projected/bb7d7a89-bc89-4b18-bbf2-626afb9452b3-kube-api-access-bw6zk\") pod \"ovn-controller-vh28h\" (UID: \"bb7d7a89-bc89-4b18-bbf2-626afb9452b3\") " pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.448222 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-var-run\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.448520 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-scripts\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.448558 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-var-lib\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.448575 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trcgx\" (UniqueName: \"kubernetes.io/projected/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-kube-api-access-trcgx\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.448612 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-etc-ovs\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.448642 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-var-log\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.448828 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-var-log\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.448875 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-var-run\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.450760 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-scripts\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.450917 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-var-lib\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.450974 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-etc-ovs\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.464989 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trcgx\" (UniqueName: \"kubernetes.io/projected/fc49c420-99e9-498c-8b80-c7d2b2c9ed18-kube-api-access-trcgx\") pod \"ovn-controller-ovs-7l6s7\" (UID: \"fc49c420-99e9-498c-8b80-c7d2b2c9ed18\") " pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.498559 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vh28h" Nov 24 14:36:09 crc kubenswrapper[4822]: I1124 14:36:09.549775 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:10 crc kubenswrapper[4822]: W1124 14:36:10.829197 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda16b9b73_6577_4420_89bb_3c774f9c4bd3.slice/crio-64f90f7e41583677d0c253bdfba33733e9fef891ccd4c174d72419db6baad9ce WatchSource:0}: Error finding container 64f90f7e41583677d0c253bdfba33733e9fef891ccd4c174d72419db6baad9ce: Status 404 returned error can't find the container with id 64f90f7e41583677d0c253bdfba33733e9fef891ccd4c174d72419db6baad9ce Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.121048 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" event={"ID":"a16b9b73-6577-4420-89bb-3c774f9c4bd3","Type":"ContainerStarted","Data":"64f90f7e41583677d0c253bdfba33733e9fef891ccd4c174d72419db6baad9ce"} Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.319933 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.502246 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.504034 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.505392 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.506505 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.506715 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.506894 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.509676 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.546938 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tkpf8" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.586055 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/890c0404-624f-4c2c-a8d8-ca9aad10c7af-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.586120 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fd303c7e-5d04-4df8-91d0-8094acef7aa7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd303c7e-5d04-4df8-91d0-8094acef7aa7\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.586158 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/890c0404-624f-4c2c-a8d8-ca9aad10c7af-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.586219 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/890c0404-624f-4c2c-a8d8-ca9aad10c7af-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.586238 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/890c0404-624f-4c2c-a8d8-ca9aad10c7af-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.586251 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/890c0404-624f-4c2c-a8d8-ca9aad10c7af-config\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.586269 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkwgr\" (UniqueName: \"kubernetes.io/projected/890c0404-624f-4c2c-a8d8-ca9aad10c7af-kube-api-access-gkwgr\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.586291 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/890c0404-624f-4c2c-a8d8-ca9aad10c7af-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.687460 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/890c0404-624f-4c2c-a8d8-ca9aad10c7af-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.687542 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/890c0404-624f-4c2c-a8d8-ca9aad10c7af-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.687563 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/890c0404-624f-4c2c-a8d8-ca9aad10c7af-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.687579 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/890c0404-624f-4c2c-a8d8-ca9aad10c7af-config\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.687598 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkwgr\" (UniqueName: \"kubernetes.io/projected/890c0404-624f-4c2c-a8d8-ca9aad10c7af-kube-api-access-gkwgr\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.687825 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/890c0404-624f-4c2c-a8d8-ca9aad10c7af-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.694447 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/890c0404-624f-4c2c-a8d8-ca9aad10c7af-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.694794 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fd303c7e-5d04-4df8-91d0-8094acef7aa7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd303c7e-5d04-4df8-91d0-8094acef7aa7\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.689695 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/890c0404-624f-4c2c-a8d8-ca9aad10c7af-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.689177 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/890c0404-624f-4c2c-a8d8-ca9aad10c7af-config\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.689470 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/890c0404-624f-4c2c-a8d8-ca9aad10c7af-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.700924 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.700967 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fd303c7e-5d04-4df8-91d0-8094acef7aa7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd303c7e-5d04-4df8-91d0-8094acef7aa7\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2251217b325d788356ff7d8451af9568f4528f781601a05dbb0016ee796ad074/globalmount\"" pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.701075 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/890c0404-624f-4c2c-a8d8-ca9aad10c7af-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.703085 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/890c0404-624f-4c2c-a8d8-ca9aad10c7af-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.707847 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkwgr\" (UniqueName: \"kubernetes.io/projected/890c0404-624f-4c2c-a8d8-ca9aad10c7af-kube-api-access-gkwgr\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.724485 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/890c0404-624f-4c2c-a8d8-ca9aad10c7af-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.746374 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fd303c7e-5d04-4df8-91d0-8094acef7aa7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd303c7e-5d04-4df8-91d0-8094acef7aa7\") pod \"ovsdbserver-nb-0\" (UID: \"890c0404-624f-4c2c-a8d8-ca9aad10c7af\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:11 crc kubenswrapper[4822]: I1124 14:36:11.870274 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: W1124 14:36:12.031279 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3571ac5_4881_4dfc_aaa2_60d7c5c39a37.slice/crio-5af2c0566f682b6183a41fec80b61b517b204d27891cd6a77a63ee67e7ff6060 WatchSource:0}: Error finding container 5af2c0566f682b6183a41fec80b61b517b204d27891cd6a77a63ee67e7ff6060: Status 404 returned error can't find the container with id 5af2c0566f682b6183a41fec80b61b517b204d27891cd6a77a63ee67e7ff6060 Nov 24 14:36:12 crc kubenswrapper[4822]: E1124 14:36:12.082353 4822 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 14:36:12 crc kubenswrapper[4822]: E1124 14:36:12.082508 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55hrz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-9mnsx_openstack(76db3a36-9f32-48e9-930d-69aee5393aa4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:36:12 crc kubenswrapper[4822]: E1124 14:36:12.083704 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" podUID="76db3a36-9f32-48e9-930d-69aee5393aa4" Nov 24 14:36:12 crc kubenswrapper[4822]: E1124 14:36:12.094308 4822 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 14:36:12 crc kubenswrapper[4822]: E1124 14:36:12.094438 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cwv8f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-q4ktn_openstack(552d1d02-3200-45d0-a913-45c5188163b9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:36:12 crc kubenswrapper[4822]: E1124 14:36:12.095551 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" podUID="552d1d02-3200-45d0-a913-45c5188163b9" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.133617 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37","Type":"ContainerStarted","Data":"5af2c0566f682b6183a41fec80b61b517b204d27891cd6a77a63ee67e7ff6060"} Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.630603 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.632281 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.633267 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.635980 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.636921 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.637283 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.637709 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-2nw5p" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.693872 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.711670 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/830f9ad0-3965-4887-b131-df42fdb35db6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.711737 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/830f9ad0-3965-4887-b131-df42fdb35db6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.711778 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/830f9ad0-3965-4887-b131-df42fdb35db6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.711817 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/830f9ad0-3965-4887-b131-df42fdb35db6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.711835 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/830f9ad0-3965-4887-b131-df42fdb35db6-config\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.711857 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c0cbd0bd-bf8a-482e-977f-65d4bb7e5e58\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0cbd0bd-bf8a-482e-977f-65d4bb7e5e58\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.711911 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmbm7\" (UniqueName: \"kubernetes.io/projected/830f9ad0-3965-4887-b131-df42fdb35db6-kube-api-access-fmbm7\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.711940 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/830f9ad0-3965-4887-b131-df42fdb35db6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.712132 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:36:12 crc kubenswrapper[4822]: W1124 14:36:12.715621 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda52d37ed_a00f_4983_88ee_023bd282038e.slice/crio-3ef59d208b26c421729e202a5dc4b21edcaa051120d52d21eedcab90fb1fafff WatchSource:0}: Error finding container 3ef59d208b26c421729e202a5dc4b21edcaa051120d52d21eedcab90fb1fafff: Status 404 returned error can't find the container with id 3ef59d208b26c421729e202a5dc4b21edcaa051120d52d21eedcab90fb1fafff Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.716528 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.737830 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vd528"] Nov 24 14:36:12 crc kubenswrapper[4822]: W1124 14:36:12.743019 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f960c65_fc33_4f09_a525_e0542ce7207c.slice/crio-a8050d59edb0e90a6a1eb3a61c82d29dc762a93953fe05a86d64bcec389a774e WatchSource:0}: Error finding container a8050d59edb0e90a6a1eb3a61c82d29dc762a93953fe05a86d64bcec389a774e: Status 404 returned error can't find the container with id a8050d59edb0e90a6a1eb3a61c82d29dc762a93953fe05a86d64bcec389a774e Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.818692 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-dns-svc\") pod \"76db3a36-9f32-48e9-930d-69aee5393aa4\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.818834 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55hrz\" (UniqueName: \"kubernetes.io/projected/76db3a36-9f32-48e9-930d-69aee5393aa4-kube-api-access-55hrz\") pod \"76db3a36-9f32-48e9-930d-69aee5393aa4\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.818866 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwv8f\" (UniqueName: \"kubernetes.io/projected/552d1d02-3200-45d0-a913-45c5188163b9-kube-api-access-cwv8f\") pod \"552d1d02-3200-45d0-a913-45c5188163b9\" (UID: \"552d1d02-3200-45d0-a913-45c5188163b9\") " Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.818895 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-config\") pod \"76db3a36-9f32-48e9-930d-69aee5393aa4\" (UID: \"76db3a36-9f32-48e9-930d-69aee5393aa4\") " Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.818952 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552d1d02-3200-45d0-a913-45c5188163b9-config\") pod \"552d1d02-3200-45d0-a913-45c5188163b9\" (UID: \"552d1d02-3200-45d0-a913-45c5188163b9\") " Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.819195 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmbm7\" (UniqueName: \"kubernetes.io/projected/830f9ad0-3965-4887-b131-df42fdb35db6-kube-api-access-fmbm7\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.819279 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/830f9ad0-3965-4887-b131-df42fdb35db6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.819354 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/830f9ad0-3965-4887-b131-df42fdb35db6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.819415 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/830f9ad0-3965-4887-b131-df42fdb35db6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.819475 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/830f9ad0-3965-4887-b131-df42fdb35db6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.819534 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/830f9ad0-3965-4887-b131-df42fdb35db6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.819559 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/830f9ad0-3965-4887-b131-df42fdb35db6-config\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.819607 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c0cbd0bd-bf8a-482e-977f-65d4bb7e5e58\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0cbd0bd-bf8a-482e-977f-65d4bb7e5e58\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.820859 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "76db3a36-9f32-48e9-930d-69aee5393aa4" (UID: "76db3a36-9f32-48e9-930d-69aee5393aa4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.823479 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/830f9ad0-3965-4887-b131-df42fdb35db6-config\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.828745 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-config" (OuterVolumeSpecName: "config") pod "76db3a36-9f32-48e9-930d-69aee5393aa4" (UID: "76db3a36-9f32-48e9-930d-69aee5393aa4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.829507 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/830f9ad0-3965-4887-b131-df42fdb35db6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.830111 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/830f9ad0-3965-4887-b131-df42fdb35db6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.830721 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/552d1d02-3200-45d0-a913-45c5188163b9-config" (OuterVolumeSpecName: "config") pod "552d1d02-3200-45d0-a913-45c5188163b9" (UID: "552d1d02-3200-45d0-a913-45c5188163b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.831001 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/830f9ad0-3965-4887-b131-df42fdb35db6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.831633 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/830f9ad0-3965-4887-b131-df42fdb35db6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.833926 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.833954 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c0cbd0bd-bf8a-482e-977f-65d4bb7e5e58\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0cbd0bd-bf8a-482e-977f-65d4bb7e5e58\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5c31b27302b2e27b6622f76a8f9b32d1bae497a3fecf4116deeb477d19077a54/globalmount\"" pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.834554 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/830f9ad0-3965-4887-b131-df42fdb35db6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.843679 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76db3a36-9f32-48e9-930d-69aee5393aa4-kube-api-access-55hrz" (OuterVolumeSpecName: "kube-api-access-55hrz") pod "76db3a36-9f32-48e9-930d-69aee5393aa4" (UID: "76db3a36-9f32-48e9-930d-69aee5393aa4"). InnerVolumeSpecName "kube-api-access-55hrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.848121 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/552d1d02-3200-45d0-a913-45c5188163b9-kube-api-access-cwv8f" (OuterVolumeSpecName: "kube-api-access-cwv8f") pod "552d1d02-3200-45d0-a913-45c5188163b9" (UID: "552d1d02-3200-45d0-a913-45c5188163b9"). InnerVolumeSpecName "kube-api-access-cwv8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.852601 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vh28h"] Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.856728 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmbm7\" (UniqueName: \"kubernetes.io/projected/830f9ad0-3965-4887-b131-df42fdb35db6-kube-api-access-fmbm7\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.876886 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.904526 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c0cbd0bd-bf8a-482e-977f-65d4bb7e5e58\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0cbd0bd-bf8a-482e-977f-65d4bb7e5e58\") pod \"ovsdbserver-sb-0\" (UID: \"830f9ad0-3965-4887-b131-df42fdb35db6\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.907880 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.916193 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.921305 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.921328 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55hrz\" (UniqueName: \"kubernetes.io/projected/76db3a36-9f32-48e9-930d-69aee5393aa4-kube-api-access-55hrz\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.921340 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwv8f\" (UniqueName: \"kubernetes.io/projected/552d1d02-3200-45d0-a913-45c5188163b9-kube-api-access-cwv8f\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.921350 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76db3a36-9f32-48e9-930d-69aee5393aa4-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.921358 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552d1d02-3200-45d0-a913-45c5188163b9-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:12 crc kubenswrapper[4822]: I1124 14:36:12.967889 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.061075 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.088267 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.101833 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.147065 4822 generic.go:334] "Generic (PLEG): container finished" podID="a16b9b73-6577-4420-89bb-3c774f9c4bd3" containerID="86dc9fbc9b39772eca9e0f1d7131deb8ee991c4d5e2fb5f2025c3b6b42025e36" exitCode=0 Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.147498 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" event={"ID":"a16b9b73-6577-4420-89bb-3c774f9c4bd3","Type":"ContainerDied","Data":"86dc9fbc9b39772eca9e0f1d7131deb8ee991c4d5e2fb5f2025c3b6b42025e36"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.173618 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a756bbea-7fc2-443e-8d92-37f993ad9ad8","Type":"ContainerStarted","Data":"e891e806e22019176a598fe59d3a5724fd4fd20e45b9ae380016fe39ee4d0f84"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.180392 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"346023fa-f08b-4ddb-a527-a637be223953","Type":"ContainerStarted","Data":"eba20e3bb59dcd6a61aaca95f841e238308ae67033123a900d614842faa5af01"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.183790 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a52d37ed-a00f-4983-88ee-023bd282038e","Type":"ContainerStarted","Data":"3ef59d208b26c421729e202a5dc4b21edcaa051120d52d21eedcab90fb1fafff"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.192679 4822 generic.go:334] "Generic (PLEG): container finished" podID="8f960c65-fc33-4f09-a525-e0542ce7207c" containerID="8bd88d4804ede327a228082bc760bbc64ceb54ce2bcdee57361364ea7ef45712" exitCode=0 Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.192821 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vd528" event={"ID":"8f960c65-fc33-4f09-a525-e0542ce7207c","Type":"ContainerDied","Data":"8bd88d4804ede327a228082bc760bbc64ceb54ce2bcdee57361364ea7ef45712"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.192859 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vd528" event={"ID":"8f960c65-fc33-4f09-a525-e0542ce7207c","Type":"ContainerStarted","Data":"a8050d59edb0e90a6a1eb3a61c82d29dc762a93953fe05a86d64bcec389a774e"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.194290 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e5326828-6e3f-464a-90d6-0cf579f5c598","Type":"ContainerStarted","Data":"dc86a6d808d6cb9bc23fdca6be620b51b537cc2d7d07a3e44eb2b01670582c3c"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.201653 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" event={"ID":"76db3a36-9f32-48e9-930d-69aee5393aa4","Type":"ContainerDied","Data":"9f1a4969cc0ea25440ae085977a30791c7cc16a96eff13b20b349206a86d258a"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.201750 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9mnsx" Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.222698 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4c617427-bd54-45c7-8e20-4ed74395e58c","Type":"ContainerStarted","Data":"15ac2f184e8e3b62b924cf888e3abf30ff56605009b5b323bca0e2cd974e1528"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.230475 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vh28h" event={"ID":"bb7d7a89-bc89-4b18-bbf2-626afb9452b3","Type":"ContainerStarted","Data":"38ff667c12cedc972a3dde68be41a3a97c686bd5a7670bbcde8a6f667364ab77"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.233550 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.240964 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42","Type":"ContainerStarted","Data":"88248f1b5ee6abb752d7444c932143936dc492a1265f1d7662850d742923df5f"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.242156 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerStarted","Data":"24f52fd79033d679211dfe7f06ada77f47d35d873445b1e5ad8fed0d10d6e535"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.245013 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" event={"ID":"552d1d02-3200-45d0-a913-45c5188163b9","Type":"ContainerDied","Data":"3d5362d26e793b20896b661855be65fb83c50336705cb8f8bca4503932a7c7cb"} Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.245073 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-q4ktn" Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.298354 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9mnsx"] Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.313111 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9mnsx"] Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.350119 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7l6s7"] Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.355894 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4ktn"] Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.358457 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4ktn"] Nov 24 14:36:13 crc kubenswrapper[4822]: E1124 14:36:13.497320 4822 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 24 14:36:13 crc kubenswrapper[4822]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/8f960c65-fc33-4f09-a525-e0542ce7207c/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 24 14:36:13 crc kubenswrapper[4822]: > podSandboxID="a8050d59edb0e90a6a1eb3a61c82d29dc762a93953fe05a86d64bcec389a774e" Nov 24 14:36:13 crc kubenswrapper[4822]: E1124 14:36:13.497453 4822 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 24 14:36:13 crc kubenswrapper[4822]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2ggpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-vd528_openstack(8f960c65-fc33-4f09-a525-e0542ce7207c): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/8f960c65-fc33-4f09-a525-e0542ce7207c/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 24 14:36:13 crc kubenswrapper[4822]: > logger="UnhandledError" Nov 24 14:36:13 crc kubenswrapper[4822]: E1124 14:36:13.498747 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/8f960c65-fc33-4f09-a525-e0542ce7207c/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-vd528" podUID="8f960c65-fc33-4f09-a525-e0542ce7207c" Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.509717 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 14:36:13 crc kubenswrapper[4822]: W1124 14:36:13.517792 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod830f9ad0_3965_4887_b131_df42fdb35db6.slice/crio-729eeab55f23f92b3150b7f856695b7e928e7330c6d15e182e985cc00222cd0b WatchSource:0}: Error finding container 729eeab55f23f92b3150b7f856695b7e928e7330c6d15e182e985cc00222cd0b: Status 404 returned error can't find the container with id 729eeab55f23f92b3150b7f856695b7e928e7330c6d15e182e985cc00222cd0b Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.735592 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="552d1d02-3200-45d0-a913-45c5188163b9" path="/var/lib/kubelet/pods/552d1d02-3200-45d0-a913-45c5188163b9/volumes" Nov 24 14:36:13 crc kubenswrapper[4822]: I1124 14:36:13.735968 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76db3a36-9f32-48e9-930d-69aee5393aa4" path="/var/lib/kubelet/pods/76db3a36-9f32-48e9-930d-69aee5393aa4/volumes" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.253944 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"830f9ad0-3965-4887-b131-df42fdb35db6","Type":"ContainerStarted","Data":"729eeab55f23f92b3150b7f856695b7e928e7330c6d15e182e985cc00222cd0b"} Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.255834 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"890c0404-624f-4c2c-a8d8-ca9aad10c7af","Type":"ContainerStarted","Data":"eea08217d814ea1cfe27fa18a96313d6012d80bde3cc528feb449a7abd391f4c"} Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.258087 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" event={"ID":"a16b9b73-6577-4420-89bb-3c774f9c4bd3","Type":"ContainerStarted","Data":"61f573942a5401933034ae3d1ba91c22291a15242f22a603ea0f4d737cacdb03"} Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.258140 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.260790 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7l6s7" event={"ID":"fc49c420-99e9-498c-8b80-c7d2b2c9ed18","Type":"ContainerStarted","Data":"25801ab35030630df312e2b328759931e291ff7c9617d257b4d3b1f4fdd413a1"} Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.295727 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" podStartSLOduration=13.857500404 podStartE2EDuration="15.295708786s" podCreationTimestamp="2025-11-24 14:35:59 +0000 UTC" firstStartedPulling="2025-11-24 14:36:10.838287464 +0000 UTC m=+1007.954927971" lastFinishedPulling="2025-11-24 14:36:12.276495876 +0000 UTC m=+1009.393136353" observedRunningTime="2025-11-24 14:36:14.275942369 +0000 UTC m=+1011.392582846" watchObservedRunningTime="2025-11-24 14:36:14.295708786 +0000 UTC m=+1011.412349263" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.879035 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv"] Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.881642 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.885673 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca-bundle" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.885862 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-http" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.886274 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-dockercfg-78gj9" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.886414 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-grpc" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.886274 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-config" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.893633 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv"] Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.969900 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.970183 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.970524 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-config\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.970711 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:14 crc kubenswrapper[4822]: I1124 14:36:14.970752 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcrtr\" (UniqueName: \"kubernetes.io/projected/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-kube-api-access-lcrtr\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.071774 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.071813 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-config\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.071867 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.071888 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcrtr\" (UniqueName: \"kubernetes.io/projected/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-kube-api-access-lcrtr\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.071936 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.074085 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp"] Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.075182 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.077253 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.078051 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-config\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.082545 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.097195 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.117552 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-loki-s3" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.117750 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-http" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.117872 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-grpc" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.121687 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp"] Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.127471 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q"] Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.128523 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.165598 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcrtr\" (UniqueName: \"kubernetes.io/projected/64db7204-1d9c-48c9-92b7-1fb1d468aa6c-kube-api-access-lcrtr\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-mvpjv\" (UID: \"64db7204-1d9c-48c9-92b7-1fb1d468aa6c\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.168650 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-grpc" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.174710 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/66287d73-c11d-48ba-aba2-570671d308e2-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.177712 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9m6q\" (UniqueName: \"kubernetes.io/projected/f03659dc-ab31-4c65-90d4-a94c4044b76e-kube-api-access-l9m6q\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.177885 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66287d73-c11d-48ba-aba2-570671d308e2-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.177943 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.177976 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/66287d73-c11d-48ba-aba2-570671d308e2-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.178004 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f03659dc-ab31-4c65-90d4-a94c4044b76e-config\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.178023 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2vj4\" (UniqueName: \"kubernetes.io/projected/66287d73-c11d-48ba-aba2-570671d308e2-kube-api-access-r2vj4\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.178076 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.178138 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.178179 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66287d73-c11d-48ba-aba2-570671d308e2-config\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.178222 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.170294 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-http" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.183032 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q"] Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.214325 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.239615 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v"] Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.240655 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.252909 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-client-http" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.252967 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.253115 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway-ca-bundle" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.253174 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.253307 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-http" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.254029 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.254278 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-dockercfg-gvlhh" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.268708 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v"] Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.277003 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq"] Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.278455 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281014 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66287d73-c11d-48ba-aba2-570671d308e2-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281071 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281094 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/66287d73-c11d-48ba-aba2-570671d308e2-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281117 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2vj4\" (UniqueName: \"kubernetes.io/projected/66287d73-c11d-48ba-aba2-570671d308e2-kube-api-access-r2vj4\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281132 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f03659dc-ab31-4c65-90d4-a94c4044b76e-config\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281159 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281189 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281280 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66287d73-c11d-48ba-aba2-570671d308e2-config\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281303 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281323 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/66287d73-c11d-48ba-aba2-570671d308e2-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.281343 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9m6q\" (UniqueName: \"kubernetes.io/projected/f03659dc-ab31-4c65-90d4-a94c4044b76e-kube-api-access-l9m6q\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.282313 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66287d73-c11d-48ba-aba2-570671d308e2-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.292100 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66287d73-c11d-48ba-aba2-570671d308e2-config\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.294877 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.298003 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f03659dc-ab31-4c65-90d4-a94c4044b76e-config\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.298280 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.314870 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.315808 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/f03659dc-ab31-4c65-90d4-a94c4044b76e-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.321088 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/66287d73-c11d-48ba-aba2-570671d308e2-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.321198 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/66287d73-c11d-48ba-aba2-570671d308e2-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.323535 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2vj4\" (UniqueName: \"kubernetes.io/projected/66287d73-c11d-48ba-aba2-570671d308e2-kube-api-access-r2vj4\") pod \"cloudkitty-lokistack-query-frontend-779849886d-f557q\" (UID: \"66287d73-c11d-48ba-aba2-570671d308e2\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.323596 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq"] Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.324327 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9m6q\" (UniqueName: \"kubernetes.io/projected/f03659dc-ab31-4c65-90d4-a94c4044b76e-kube-api-access-l9m6q\") pod \"cloudkitty-lokistack-querier-548665d79b-fjkbp\" (UID: \"f03659dc-ab31-4c65-90d4-a94c4044b76e\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383413 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/b1986632-6691-4479-a0b5-a7dc71eda136-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383499 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6z7c\" (UniqueName: \"kubernetes.io/projected/b9bc089e-4276-4b2d-b7b3-8f455629074c-kube-api-access-b6z7c\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383523 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383547 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383594 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383665 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/b1986632-6691-4479-a0b5-a7dc71eda136-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383732 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383765 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383788 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383810 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/b9bc089e-4276-4b2d-b7b3-8f455629074c-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.383828 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.384410 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79r5c\" (UniqueName: \"kubernetes.io/projected/b1986632-6691-4479-a0b5-a7dc71eda136-kube-api-access-79r5c\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.384520 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.384542 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.384585 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.384606 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/b9bc089e-4276-4b2d-b7b3-8f455629074c-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.384624 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.384655 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493652 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493716 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493741 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493767 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/b9bc089e-4276-4b2d-b7b3-8f455629074c-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493784 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493825 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79r5c\" (UniqueName: \"kubernetes.io/projected/b1986632-6691-4479-a0b5-a7dc71eda136-kube-api-access-79r5c\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493855 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493879 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493909 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493930 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/b9bc089e-4276-4b2d-b7b3-8f455629074c-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493944 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493972 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.493993 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/b1986632-6691-4479-a0b5-a7dc71eda136-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.494015 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6z7c\" (UniqueName: \"kubernetes.io/projected/b9bc089e-4276-4b2d-b7b3-8f455629074c-kube-api-access-b6z7c\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.494029 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.494047 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.494075 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.494103 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/b1986632-6691-4479-a0b5-a7dc71eda136-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.494700 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.496019 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.496289 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.497175 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.497889 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.497914 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.498172 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.498260 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.501433 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/b9bc089e-4276-4b2d-b7b3-8f455629074c-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.501804 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/b9bc089e-4276-4b2d-b7b3-8f455629074c-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.501921 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.503881 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/b9bc089e-4276-4b2d-b7b3-8f455629074c-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.503910 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/b9bc089e-4276-4b2d-b7b3-8f455629074c-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.503951 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/b1986632-6691-4479-a0b5-a7dc71eda136-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.504403 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/b1986632-6691-4479-a0b5-a7dc71eda136-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.505831 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/b1986632-6691-4479-a0b5-a7dc71eda136-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.509516 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.512675 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79r5c\" (UniqueName: \"kubernetes.io/projected/b1986632-6691-4479-a0b5-a7dc71eda136-kube-api-access-79r5c\") pod \"cloudkitty-lokistack-gateway-76cc998948-s2zzq\" (UID: \"b1986632-6691-4479-a0b5-a7dc71eda136\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.514233 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6z7c\" (UniqueName: \"kubernetes.io/projected/b9bc089e-4276-4b2d-b7b3-8f455629074c-kube-api-access-b6z7c\") pod \"cloudkitty-lokistack-gateway-76cc998948-mz99v\" (UID: \"b9bc089e-4276-4b2d-b7b3-8f455629074c\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.520019 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.583993 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:15 crc kubenswrapper[4822]: I1124 14:36:15.667863 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.051805 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.053011 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.055994 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-grpc" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.063990 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-http" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.086255 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.108396 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/849f7f3c-3605-4e1b-8044-5e063841745e-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.108450 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.108531 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.108601 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.108666 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.108722 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29nsj\" (UniqueName: \"kubernetes.io/projected/849f7f3c-3605-4e1b-8044-5e063841745e-kube-api-access-29nsj\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.108751 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.108770 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.125922 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.128683 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.134887 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-grpc" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.137470 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-http" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.140419 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.177723 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.179382 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.184878 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-http" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.184889 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-grpc" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.194082 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.215011 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/849f7f3c-3605-4e1b-8044-5e063841745e-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.215331 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.215355 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216023 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/849f7f3c-3605-4e1b-8044-5e063841745e-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216027 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216119 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v726q\" (UniqueName: \"kubernetes.io/projected/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-kube-api-access-v726q\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216177 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216235 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216297 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216342 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216363 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216393 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216417 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216440 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216464 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216484 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216518 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29nsj\" (UniqueName: \"kubernetes.io/projected/849f7f3c-3605-4e1b-8044-5e063841745e-kube-api-access-29nsj\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216539 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216559 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216575 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216594 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216620 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzwsj\" (UniqueName: \"kubernetes.io/projected/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-kube-api-access-wzwsj\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216642 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.216846 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.217761 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.218294 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.220640 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.231999 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.240391 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.242483 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/849f7f3c-3605-4e1b-8044-5e063841745e-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.245902 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.253237 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29nsj\" (UniqueName: \"kubernetes.io/projected/849f7f3c-3605-4e1b-8044-5e063841745e-kube-api-access-29nsj\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"849f7f3c-3605-4e1b-8044-5e063841745e\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318244 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318294 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzwsj\" (UniqueName: \"kubernetes.io/projected/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-kube-api-access-wzwsj\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318331 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318357 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318406 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318428 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v726q\" (UniqueName: \"kubernetes.io/projected/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-kube-api-access-v726q\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318456 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318505 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318527 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318568 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318588 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318608 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318647 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.318680 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.319645 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.320374 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.321016 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.321454 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.321905 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.322134 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.325562 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.326455 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.328631 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.329369 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.329446 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.337836 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.338105 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzwsj\" (UniqueName: \"kubernetes.io/projected/8f57fdf8-920a-4329-a0b1-89b62cfa82a6-kube-api-access-wzwsj\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.354183 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v726q\" (UniqueName: \"kubernetes.io/projected/ac81e91d-a816-4a86-98b6-3e43f36c7b6f-kube-api-access-v726q\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.361425 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8f57fdf8-920a-4329-a0b1-89b62cfa82a6\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.378165 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.379135 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"ac81e91d-a816-4a86-98b6-3e43f36c7b6f\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.450995 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:16 crc kubenswrapper[4822]: I1124 14:36:16.500791 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:19 crc kubenswrapper[4822]: I1124 14:36:19.594377 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:36:19 crc kubenswrapper[4822]: I1124 14:36:19.650557 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vd528"] Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.010747 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-n9zjj"] Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.013420 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.017522 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.029287 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-n9zjj"] Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.116824 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.116878 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqvd5\" (UniqueName: \"kubernetes.io/projected/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-kube-api-access-vqvd5\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.116896 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-ovs-rundir\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.116919 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-config\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.116973 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-ovn-rundir\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.117010 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-combined-ca-bundle\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.218138 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.218199 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqvd5\" (UniqueName: \"kubernetes.io/projected/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-kube-api-access-vqvd5\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.218236 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-ovs-rundir\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.218259 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-config\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.218300 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-ovn-rundir\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.218342 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-combined-ca-bundle\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.218561 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-ovs-rundir\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.218868 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-ovn-rundir\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.219155 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-config\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.223707 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.223770 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-combined-ca-bundle\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.249738 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqvd5\" (UniqueName: \"kubernetes.io/projected/e6c5c30b-4d09-4d55-b483-a6b1a6cfee86-kube-api-access-vqvd5\") pod \"ovn-controller-metrics-n9zjj\" (UID: \"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86\") " pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.255710 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxxdv"] Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.262075 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.265548 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.278380 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxxdv"] Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.320425 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdzbt\" (UniqueName: \"kubernetes.io/projected/c9defc3c-923f-469f-9d93-0902ce97aa3d-kube-api-access-bdzbt\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.320472 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.320502 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-config\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.320570 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.350871 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-n9zjj" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.412709 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxxdv"] Nov 24 14:36:21 crc kubenswrapper[4822]: E1124 14:36:21.413386 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-bdzbt ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" podUID="c9defc3c-923f-469f-9d93-0902ce97aa3d" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.422365 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdzbt\" (UniqueName: \"kubernetes.io/projected/c9defc3c-923f-469f-9d93-0902ce97aa3d-kube-api-access-bdzbt\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.422417 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.422441 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-config\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.422489 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.423440 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.423602 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.424004 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-config\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.435452 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7ccfg"] Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.436858 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.462105 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdzbt\" (UniqueName: \"kubernetes.io/projected/c9defc3c-923f-469f-9d93-0902ce97aa3d-kube-api-access-bdzbt\") pod \"dnsmasq-dns-7fd796d7df-pxxdv\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.462431 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.472233 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7ccfg"] Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.523770 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctjnf\" (UniqueName: \"kubernetes.io/projected/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-kube-api-access-ctjnf\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.523910 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.523947 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-config\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.523971 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.524062 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.625370 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-config\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.625424 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.625547 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.625621 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctjnf\" (UniqueName: \"kubernetes.io/projected/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-kube-api-access-ctjnf\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.626721 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.626729 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.626886 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.625730 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.627175 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-config\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.648388 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctjnf\" (UniqueName: \"kubernetes.io/projected/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-kube-api-access-ctjnf\") pod \"dnsmasq-dns-86db49b7ff-7ccfg\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.798179 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:21 crc kubenswrapper[4822]: I1124 14:36:21.962406 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v"] Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.323600 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.334352 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.442268 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-dns-svc\") pod \"c9defc3c-923f-469f-9d93-0902ce97aa3d\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.442338 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-config\") pod \"c9defc3c-923f-469f-9d93-0902ce97aa3d\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.442520 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdzbt\" (UniqueName: \"kubernetes.io/projected/c9defc3c-923f-469f-9d93-0902ce97aa3d-kube-api-access-bdzbt\") pod \"c9defc3c-923f-469f-9d93-0902ce97aa3d\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.442555 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-ovsdbserver-nb\") pod \"c9defc3c-923f-469f-9d93-0902ce97aa3d\" (UID: \"c9defc3c-923f-469f-9d93-0902ce97aa3d\") " Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.443020 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-config" (OuterVolumeSpecName: "config") pod "c9defc3c-923f-469f-9d93-0902ce97aa3d" (UID: "c9defc3c-923f-469f-9d93-0902ce97aa3d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.443120 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9defc3c-923f-469f-9d93-0902ce97aa3d" (UID: "c9defc3c-923f-469f-9d93-0902ce97aa3d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.443670 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9defc3c-923f-469f-9d93-0902ce97aa3d" (UID: "c9defc3c-923f-469f-9d93-0902ce97aa3d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.454289 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9defc3c-923f-469f-9d93-0902ce97aa3d-kube-api-access-bdzbt" (OuterVolumeSpecName: "kube-api-access-bdzbt") pod "c9defc3c-923f-469f-9d93-0902ce97aa3d" (UID: "c9defc3c-923f-469f-9d93-0902ce97aa3d"). InnerVolumeSpecName "kube-api-access-bdzbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.545278 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.545315 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.545327 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdzbt\" (UniqueName: \"kubernetes.io/projected/c9defc3c-923f-469f-9d93-0902ce97aa3d-kube-api-access-bdzbt\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:22 crc kubenswrapper[4822]: I1124 14:36:22.545339 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9defc3c-923f-469f-9d93-0902ce97aa3d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:23 crc kubenswrapper[4822]: I1124 14:36:23.330895 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-pxxdv" Nov 24 14:36:23 crc kubenswrapper[4822]: I1124 14:36:23.405270 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxxdv"] Nov 24 14:36:23 crc kubenswrapper[4822]: I1124 14:36:23.416345 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxxdv"] Nov 24 14:36:23 crc kubenswrapper[4822]: W1124 14:36:23.573710 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9bc089e_4276_4b2d_b7b3_8f455629074c.slice/crio-7246c7064882a336d733330a85f21615825e70b29cd953a20ce5da554b3aec5b WatchSource:0}: Error finding container 7246c7064882a336d733330a85f21615825e70b29cd953a20ce5da554b3aec5b: Status 404 returned error can't find the container with id 7246c7064882a336d733330a85f21615825e70b29cd953a20ce5da554b3aec5b Nov 24 14:36:23 crc kubenswrapper[4822]: I1124 14:36:23.738006 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9defc3c-923f-469f-9d93-0902ce97aa3d" path="/var/lib/kubelet/pods/c9defc3c-923f-469f-9d93-0902ce97aa3d/volumes" Nov 24 14:36:23 crc kubenswrapper[4822]: I1124 14:36:23.990330 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv"] Nov 24 14:36:24 crc kubenswrapper[4822]: I1124 14:36:24.331620 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp"] Nov 24 14:36:24 crc kubenswrapper[4822]: I1124 14:36:24.338226 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" event={"ID":"b9bc089e-4276-4b2d-b7b3-8f455629074c","Type":"ContainerStarted","Data":"7246c7064882a336d733330a85f21615825e70b29cd953a20ce5da554b3aec5b"} Nov 24 14:36:24 crc kubenswrapper[4822]: I1124 14:36:24.355165 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Nov 24 14:36:24 crc kubenswrapper[4822]: W1124 14:36:24.444591 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db7204_1d9c_48c9_92b7_1fb1d468aa6c.slice/crio-958df750f1fb1651104a5d6493bad0a96877be1df750b76497f692265f81da7b WatchSource:0}: Error finding container 958df750f1fb1651104a5d6493bad0a96877be1df750b76497f692265f81da7b: Status 404 returned error can't find the container with id 958df750f1fb1651104a5d6493bad0a96877be1df750b76497f692265f81da7b Nov 24 14:36:24 crc kubenswrapper[4822]: W1124 14:36:24.456070 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf03659dc_ab31_4c65_90d4_a94c4044b76e.slice/crio-b75dfdafbca06c263c1a2c54893a1ebe33db0a69e308d6107c95ea28dc4baf9c WatchSource:0}: Error finding container b75dfdafbca06c263c1a2c54893a1ebe33db0a69e308d6107c95ea28dc4baf9c: Status 404 returned error can't find the container with id b75dfdafbca06c263c1a2c54893a1ebe33db0a69e308d6107c95ea28dc4baf9c Nov 24 14:36:24 crc kubenswrapper[4822]: I1124 14:36:24.583987 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-n9zjj"] Nov 24 14:36:24 crc kubenswrapper[4822]: I1124 14:36:24.594364 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Nov 24 14:36:24 crc kubenswrapper[4822]: W1124 14:36:24.599711 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66287d73_c11d_48ba_aba2_570671d308e2.slice/crio-8b24971cb5bc164889ff90b91a44c48ed6c3f6db38f53a003febd4a117a4b00b WatchSource:0}: Error finding container 8b24971cb5bc164889ff90b91a44c48ed6c3f6db38f53a003febd4a117a4b00b: Status 404 returned error can't find the container with id 8b24971cb5bc164889ff90b91a44c48ed6c3f6db38f53a003febd4a117a4b00b Nov 24 14:36:24 crc kubenswrapper[4822]: W1124 14:36:24.601323 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1986632_6691_4479_a0b5_a7dc71eda136.slice/crio-4817b93a3ec6d9915247ea03b9fa90cc6a33809873c543d96c63f73a661bca2e WatchSource:0}: Error finding container 4817b93a3ec6d9915247ea03b9fa90cc6a33809873c543d96c63f73a661bca2e: Status 404 returned error can't find the container with id 4817b93a3ec6d9915247ea03b9fa90cc6a33809873c543d96c63f73a661bca2e Nov 24 14:36:24 crc kubenswrapper[4822]: I1124 14:36:24.602446 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q"] Nov 24 14:36:24 crc kubenswrapper[4822]: W1124 14:36:24.604602 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6c5c30b_4d09_4d55_b483_a6b1a6cfee86.slice/crio-8642632efaf9628fbe200707d7d8e9704f772c80c092ce84074eeb6f171b4e00 WatchSource:0}: Error finding container 8642632efaf9628fbe200707d7d8e9704f772c80c092ce84074eeb6f171b4e00: Status 404 returned error can't find the container with id 8642632efaf9628fbe200707d7d8e9704f772c80c092ce84074eeb6f171b4e00 Nov 24 14:36:24 crc kubenswrapper[4822]: W1124 14:36:24.610307 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f57fdf8_920a_4329_a0b1_89b62cfa82a6.slice/crio-95c03c30f68d3e4b617d751c7f2134db21d415e4fdf3bc0600ee238b86759d45 WatchSource:0}: Error finding container 95c03c30f68d3e4b617d751c7f2134db21d415e4fdf3bc0600ee238b86759d45: Status 404 returned error can't find the container with id 95c03c30f68d3e4b617d751c7f2134db21d415e4fdf3bc0600ee238b86759d45 Nov 24 14:36:24 crc kubenswrapper[4822]: I1124 14:36:24.612081 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq"] Nov 24 14:36:24 crc kubenswrapper[4822]: W1124 14:36:24.616063 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod849f7f3c_3605_4e1b_8044_5e063841745e.slice/crio-be3fb192d509c0492d133423168a9d85afd6e5e576d123ac45b0eee0bae54d9f WatchSource:0}: Error finding container be3fb192d509c0492d133423168a9d85afd6e5e576d123ac45b0eee0bae54d9f: Status 404 returned error can't find the container with id be3fb192d509c0492d133423168a9d85afd6e5e576d123ac45b0eee0bae54d9f Nov 24 14:36:24 crc kubenswrapper[4822]: I1124 14:36:24.619037 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Nov 24 14:36:24 crc kubenswrapper[4822]: I1124 14:36:24.960161 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7ccfg"] Nov 24 14:36:24 crc kubenswrapper[4822]: W1124 14:36:24.999270 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67610eb7_59cf_4ee6_bc07_0645cd3f41d1.slice/crio-700bf04e145b17cd7009a3bc80ef630ace650c083763b935ced7b55341778916 WatchSource:0}: Error finding container 700bf04e145b17cd7009a3bc80ef630ace650c083763b935ced7b55341778916: Status 404 returned error can't find the container with id 700bf04e145b17cd7009a3bc80ef630ace650c083763b935ced7b55341778916 Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.352762 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"830f9ad0-3965-4887-b131-df42fdb35db6","Type":"ContainerStarted","Data":"44722db81c3adf8cf1bb1a69a3d66e55e7ec53edaaaf9199a2b3f716589c8bcd"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.355461 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" event={"ID":"b1986632-6691-4479-a0b5-a7dc71eda136","Type":"ContainerStarted","Data":"4817b93a3ec6d9915247ea03b9fa90cc6a33809873c543d96c63f73a661bca2e"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.357027 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a756bbea-7fc2-443e-8d92-37f993ad9ad8","Type":"ContainerStarted","Data":"7a83f8e6093e39260518a1c257bf52d0142193c3f2317bfeaaf0099a590ff520"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.360560 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"ac81e91d-a816-4a86-98b6-3e43f36c7b6f","Type":"ContainerStarted","Data":"20184e4e8fca182a5ca44d26ef43ff770dabd194ea16ac225df861c10d51797b"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.362560 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" event={"ID":"f03659dc-ab31-4c65-90d4-a94c4044b76e","Type":"ContainerStarted","Data":"b75dfdafbca06c263c1a2c54893a1ebe33db0a69e308d6107c95ea28dc4baf9c"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.364161 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" event={"ID":"66287d73-c11d-48ba-aba2-570671d308e2","Type":"ContainerStarted","Data":"8b24971cb5bc164889ff90b91a44c48ed6c3f6db38f53a003febd4a117a4b00b"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.365899 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"849f7f3c-3605-4e1b-8044-5e063841745e","Type":"ContainerStarted","Data":"be3fb192d509c0492d133423168a9d85afd6e5e576d123ac45b0eee0bae54d9f"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.367287 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" event={"ID":"67610eb7-59cf-4ee6-bc07-0645cd3f41d1","Type":"ContainerStarted","Data":"700bf04e145b17cd7009a3bc80ef630ace650c083763b935ced7b55341778916"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.369400 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"8f57fdf8-920a-4329-a0b1-89b62cfa82a6","Type":"ContainerStarted","Data":"95c03c30f68d3e4b617d751c7f2134db21d415e4fdf3bc0600ee238b86759d45"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.374901 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"346023fa-f08b-4ddb-a527-a637be223953","Type":"ContainerStarted","Data":"480016c926c4ab532960a4fe6f5cadbf2e66845d0e8f9fb50ba256c3393be65c"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.375000 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.376322 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" event={"ID":"64db7204-1d9c-48c9-92b7-1fb1d468aa6c","Type":"ContainerStarted","Data":"958df750f1fb1651104a5d6493bad0a96877be1df750b76497f692265f81da7b"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.377720 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-n9zjj" event={"ID":"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86","Type":"ContainerStarted","Data":"8642632efaf9628fbe200707d7d8e9704f772c80c092ce84074eeb6f171b4e00"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.381938 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vd528" event={"ID":"8f960c65-fc33-4f09-a525-e0542ce7207c","Type":"ContainerStarted","Data":"21835380714c9cbcba6c8677e016f4c511748a6456dd539a1b8d0b8e22f07f16"} Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.382190 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-vd528" podUID="8f960c65-fc33-4f09-a525-e0542ce7207c" containerName="dnsmasq-dns" containerID="cri-o://21835380714c9cbcba6c8677e016f4c511748a6456dd539a1b8d0b8e22f07f16" gracePeriod=10 Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.382371 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:36:25 crc kubenswrapper[4822]: I1124 14:36:25.418053 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-vd528" podStartSLOduration=27.418020562 podStartE2EDuration="27.418020562s" podCreationTimestamp="2025-11-24 14:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:36:25.416630669 +0000 UTC m=+1022.533271166" watchObservedRunningTime="2025-11-24 14:36:25.418020562 +0000 UTC m=+1022.534661039" Nov 24 14:36:26 crc kubenswrapper[4822]: I1124 14:36:26.394017 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4c617427-bd54-45c7-8e20-4ed74395e58c","Type":"ContainerStarted","Data":"8e681e3638593d9d216e54ba7585d1b843a0323c02865421bd7cd35797ff81ec"} Nov 24 14:36:26 crc kubenswrapper[4822]: I1124 14:36:26.397440 4822 generic.go:334] "Generic (PLEG): container finished" podID="8f960c65-fc33-4f09-a525-e0542ce7207c" containerID="21835380714c9cbcba6c8677e016f4c511748a6456dd539a1b8d0b8e22f07f16" exitCode=0 Nov 24 14:36:26 crc kubenswrapper[4822]: I1124 14:36:26.398002 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vd528" event={"ID":"8f960c65-fc33-4f09-a525-e0542ce7207c","Type":"ContainerDied","Data":"21835380714c9cbcba6c8677e016f4c511748a6456dd539a1b8d0b8e22f07f16"} Nov 24 14:36:26 crc kubenswrapper[4822]: I1124 14:36:26.427775 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=11.860678632 podStartE2EDuration="23.427754886s" podCreationTimestamp="2025-11-24 14:36:03 +0000 UTC" firstStartedPulling="2025-11-24 14:36:12.894046578 +0000 UTC m=+1010.010687055" lastFinishedPulling="2025-11-24 14:36:24.461122832 +0000 UTC m=+1021.577763309" observedRunningTime="2025-11-24 14:36:25.449016136 +0000 UTC m=+1022.565656633" watchObservedRunningTime="2025-11-24 14:36:26.427754886 +0000 UTC m=+1023.544395373" Nov 24 14:36:27 crc kubenswrapper[4822]: I1124 14:36:27.412048 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vd528" event={"ID":"8f960c65-fc33-4f09-a525-e0542ce7207c","Type":"ContainerDied","Data":"a8050d59edb0e90a6a1eb3a61c82d29dc762a93953fe05a86d64bcec389a774e"} Nov 24 14:36:27 crc kubenswrapper[4822]: I1124 14:36:27.412493 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8050d59edb0e90a6a1eb3a61c82d29dc762a93953fe05a86d64bcec389a774e" Nov 24 14:36:27 crc kubenswrapper[4822]: I1124 14:36:27.887074 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:36:27 crc kubenswrapper[4822]: I1124 14:36:27.951371 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ggpf\" (UniqueName: \"kubernetes.io/projected/8f960c65-fc33-4f09-a525-e0542ce7207c-kube-api-access-2ggpf\") pod \"8f960c65-fc33-4f09-a525-e0542ce7207c\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " Nov 24 14:36:27 crc kubenswrapper[4822]: I1124 14:36:27.951461 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-config\") pod \"8f960c65-fc33-4f09-a525-e0542ce7207c\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " Nov 24 14:36:27 crc kubenswrapper[4822]: I1124 14:36:27.951647 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-dns-svc\") pod \"8f960c65-fc33-4f09-a525-e0542ce7207c\" (UID: \"8f960c65-fc33-4f09-a525-e0542ce7207c\") " Nov 24 14:36:28 crc kubenswrapper[4822]: I1124 14:36:28.180901 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f960c65-fc33-4f09-a525-e0542ce7207c-kube-api-access-2ggpf" (OuterVolumeSpecName: "kube-api-access-2ggpf") pod "8f960c65-fc33-4f09-a525-e0542ce7207c" (UID: "8f960c65-fc33-4f09-a525-e0542ce7207c"). InnerVolumeSpecName "kube-api-access-2ggpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:28 crc kubenswrapper[4822]: I1124 14:36:28.258724 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ggpf\" (UniqueName: \"kubernetes.io/projected/8f960c65-fc33-4f09-a525-e0542ce7207c-kube-api-access-2ggpf\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:28 crc kubenswrapper[4822]: I1124 14:36:28.316589 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-config" (OuterVolumeSpecName: "config") pod "8f960c65-fc33-4f09-a525-e0542ce7207c" (UID: "8f960c65-fc33-4f09-a525-e0542ce7207c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:28 crc kubenswrapper[4822]: I1124 14:36:28.317489 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8f960c65-fc33-4f09-a525-e0542ce7207c" (UID: "8f960c65-fc33-4f09-a525-e0542ce7207c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:28 crc kubenswrapper[4822]: I1124 14:36:28.361279 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:28 crc kubenswrapper[4822]: I1124 14:36:28.361313 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f960c65-fc33-4f09-a525-e0542ce7207c-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:28 crc kubenswrapper[4822]: I1124 14:36:28.423647 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vd528" Nov 24 14:36:28 crc kubenswrapper[4822]: I1124 14:36:28.501397 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vd528"] Nov 24 14:36:28 crc kubenswrapper[4822]: I1124 14:36:28.513185 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vd528"] Nov 24 14:36:28 crc kubenswrapper[4822]: E1124 14:36:28.875801 4822 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc49c420_99e9_498c_8b80_c7d2b2c9ed18.slice/crio-ccf1901983077b84de170f0d04dbd900d29cc2ba6fe477e969c53fa76ea7452c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc49c420_99e9_498c_8b80_c7d2b2c9ed18.slice/crio-conmon-ccf1901983077b84de170f0d04dbd900d29cc2ba6fe477e969c53fa76ea7452c.scope\": RecentStats: unable to find data in memory cache]" Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.433281 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42","Type":"ContainerStarted","Data":"5fca1ba42e03886e600ab4c599c29b43c6ef9b26c750f6df2bb1976d457df699"} Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.437153 4822 generic.go:334] "Generic (PLEG): container finished" podID="fc49c420-99e9-498c-8b80-c7d2b2c9ed18" containerID="ccf1901983077b84de170f0d04dbd900d29cc2ba6fe477e969c53fa76ea7452c" exitCode=0 Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.437247 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7l6s7" event={"ID":"fc49c420-99e9-498c-8b80-c7d2b2c9ed18","Type":"ContainerDied","Data":"ccf1901983077b84de170f0d04dbd900d29cc2ba6fe477e969c53fa76ea7452c"} Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.439962 4822 generic.go:334] "Generic (PLEG): container finished" podID="67610eb7-59cf-4ee6-bc07-0645cd3f41d1" containerID="7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd" exitCode=0 Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.440042 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" event={"ID":"67610eb7-59cf-4ee6-bc07-0645cd3f41d1","Type":"ContainerDied","Data":"7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd"} Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.453901 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a52d37ed-a00f-4983-88ee-023bd282038e","Type":"ContainerStarted","Data":"276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe"} Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.467335 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37","Type":"ContainerStarted","Data":"a11d1a87ad59947fd0c7dafe5c9ace40accaa87c88506ddfd156b36fa2cfb257"} Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.470403 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"890c0404-624f-4c2c-a8d8-ca9aad10c7af","Type":"ContainerStarted","Data":"db89448cdaf64abe8966e01ccd66befa189d0851cbd06d2da392e3d44f34cf02"} Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.471919 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vh28h" event={"ID":"bb7d7a89-bc89-4b18-bbf2-626afb9452b3","Type":"ContainerStarted","Data":"67112dee7ca5674d3898059c1f244134796456774cd9dbd2b39e74abaf52b58f"} Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.472085 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-vh28h" Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.474487 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerStarted","Data":"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539"} Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.477498 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e5326828-6e3f-464a-90d6-0cf579f5c598","Type":"ContainerStarted","Data":"7112ed58314143939d290e79da2a1799438f73664f74b747fe08285adc029d52"} Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.477768 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.544001 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vh28h" podStartSLOduration=8.860876593 podStartE2EDuration="20.543985596s" podCreationTimestamp="2025-11-24 14:36:09 +0000 UTC" firstStartedPulling="2025-11-24 14:36:12.865354266 +0000 UTC m=+1009.981994743" lastFinishedPulling="2025-11-24 14:36:24.548463269 +0000 UTC m=+1021.665103746" observedRunningTime="2025-11-24 14:36:29.542685825 +0000 UTC m=+1026.659326322" watchObservedRunningTime="2025-11-24 14:36:29.543985596 +0000 UTC m=+1026.660626073" Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.609790 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.054716944 podStartE2EDuration="24.609771018s" podCreationTimestamp="2025-11-24 14:36:05 +0000 UTC" firstStartedPulling="2025-11-24 14:36:13.082707741 +0000 UTC m=+1010.199348218" lastFinishedPulling="2025-11-24 14:36:24.637761815 +0000 UTC m=+1021.754402292" observedRunningTime="2025-11-24 14:36:29.608121528 +0000 UTC m=+1026.724762015" watchObservedRunningTime="2025-11-24 14:36:29.609771018 +0000 UTC m=+1026.726411485" Nov 24 14:36:29 crc kubenswrapper[4822]: I1124 14:36:29.716895 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f960c65-fc33-4f09-a525-e0542ce7207c" path="/var/lib/kubelet/pods/8f960c65-fc33-4f09-a525-e0542ce7207c/volumes" Nov 24 14:36:32 crc kubenswrapper[4822]: I1124 14:36:32.504959 4822 generic.go:334] "Generic (PLEG): container finished" podID="4c617427-bd54-45c7-8e20-4ed74395e58c" containerID="8e681e3638593d9d216e54ba7585d1b843a0323c02865421bd7cd35797ff81ec" exitCode=0 Nov 24 14:36:32 crc kubenswrapper[4822]: I1124 14:36:32.505055 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4c617427-bd54-45c7-8e20-4ed74395e58c","Type":"ContainerDied","Data":"8e681e3638593d9d216e54ba7585d1b843a0323c02865421bd7cd35797ff81ec"} Nov 24 14:36:32 crc kubenswrapper[4822]: I1124 14:36:32.507807 4822 generic.go:334] "Generic (PLEG): container finished" podID="a756bbea-7fc2-443e-8d92-37f993ad9ad8" containerID="7a83f8e6093e39260518a1c257bf52d0142193c3f2317bfeaaf0099a590ff520" exitCode=0 Nov 24 14:36:32 crc kubenswrapper[4822]: I1124 14:36:32.507843 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a756bbea-7fc2-443e-8d92-37f993ad9ad8","Type":"ContainerDied","Data":"7a83f8e6093e39260518a1c257bf52d0142193c3f2317bfeaaf0099a590ff520"} Nov 24 14:36:33 crc kubenswrapper[4822]: I1124 14:36:33.593571 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.529026 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" event={"ID":"67610eb7-59cf-4ee6-bc07-0645cd3f41d1","Type":"ContainerStarted","Data":"18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c"} Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.529637 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.534759 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" event={"ID":"b1986632-6691-4479-a0b5-a7dc71eda136","Type":"ContainerStarted","Data":"70d751ab9a86b695190634c0e806097e19fccb81e10c75f351f95cb9b30cb59f"} Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.535497 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.540582 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"849f7f3c-3605-4e1b-8044-5e063841745e","Type":"ContainerStarted","Data":"b9d50082b0638bcccdf1ee31c55e7cedee3e947bd278c5af6013e14f74e85b47"} Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.541105 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.549417 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4c617427-bd54-45c7-8e20-4ed74395e58c","Type":"ContainerStarted","Data":"f5be7cace6bbb3e04e948f0476aef34ce4a14167ea1565bc3ab939f56b9546a8"} Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.551305 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.552945 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a756bbea-7fc2-443e-8d92-37f993ad9ad8","Type":"ContainerStarted","Data":"eb00333156300a6ed692d6e4ada31d015aa5eb21d5a80c79d3ca0d03e3149230"} Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.554778 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" podStartSLOduration=13.554763451 podStartE2EDuration="13.554763451s" podCreationTimestamp="2025-11-24 14:36:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:36:34.550743707 +0000 UTC m=+1031.667384204" watchObservedRunningTime="2025-11-24 14:36:34.554763451 +0000 UTC m=+1031.671403928" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.554913 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7l6s7" event={"ID":"fc49c420-99e9-498c-8b80-c7d2b2c9ed18","Type":"ContainerStarted","Data":"ec8a5f5cedbc9d3f947b72fd2759d6c8c71d4e9ef013f66c84ab21052e51a559"} Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.556729 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" event={"ID":"b9bc089e-4276-4b2d-b7b3-8f455629074c","Type":"ContainerStarted","Data":"e95c165ad6b96a83e46a9dbdd3dbd6da336a335e864de98fd6fcc083281a3c50"} Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.557285 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.575481 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-ingester-0" podStartSLOduration=11.541241062 podStartE2EDuration="20.575466148s" podCreationTimestamp="2025-11-24 14:36:14 +0000 UTC" firstStartedPulling="2025-11-24 14:36:24.63271704 +0000 UTC m=+1021.749357517" lastFinishedPulling="2025-11-24 14:36:33.666942126 +0000 UTC m=+1030.783582603" observedRunningTime="2025-11-24 14:36:34.568411561 +0000 UTC m=+1031.685052048" watchObservedRunningTime="2025-11-24 14:36:34.575466148 +0000 UTC m=+1031.692106625" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.608346 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.612955 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=22.713666584 podStartE2EDuration="33.6129417s" podCreationTimestamp="2025-11-24 14:36:01 +0000 UTC" firstStartedPulling="2025-11-24 14:36:12.87814365 +0000 UTC m=+1009.994784137" lastFinishedPulling="2025-11-24 14:36:23.777418786 +0000 UTC m=+1020.894059253" observedRunningTime="2025-11-24 14:36:34.608544886 +0000 UTC m=+1031.725185363" watchObservedRunningTime="2025-11-24 14:36:34.6129417 +0000 UTC m=+1031.729582167" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.630029 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-s2zzq" podStartSLOduration=10.59740769 podStartE2EDuration="19.630013156s" podCreationTimestamp="2025-11-24 14:36:15 +0000 UTC" firstStartedPulling="2025-11-24 14:36:24.603671987 +0000 UTC m=+1021.720312464" lastFinishedPulling="2025-11-24 14:36:33.636277433 +0000 UTC m=+1030.752917930" observedRunningTime="2025-11-24 14:36:34.625619001 +0000 UTC m=+1031.742259488" watchObservedRunningTime="2025-11-24 14:36:34.630013156 +0000 UTC m=+1031.746653633" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.652940 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=23.729807561 podStartE2EDuration="34.65292085s" podCreationTimestamp="2025-11-24 14:36:00 +0000 UTC" firstStartedPulling="2025-11-24 14:36:12.894071869 +0000 UTC m=+1010.010712356" lastFinishedPulling="2025-11-24 14:36:23.817185158 +0000 UTC m=+1020.933825645" observedRunningTime="2025-11-24 14:36:34.646666178 +0000 UTC m=+1031.763306655" watchObservedRunningTime="2025-11-24 14:36:34.65292085 +0000 UTC m=+1031.769561327" Nov 24 14:36:34 crc kubenswrapper[4822]: I1124 14:36:34.686137 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mz99v" podStartSLOduration=9.640167739 podStartE2EDuration="19.686121951s" podCreationTimestamp="2025-11-24 14:36:15 +0000 UTC" firstStartedPulling="2025-11-24 14:36:23.599634748 +0000 UTC m=+1020.716275235" lastFinishedPulling="2025-11-24 14:36:33.64558897 +0000 UTC m=+1030.762229447" observedRunningTime="2025-11-24 14:36:34.680066465 +0000 UTC m=+1031.796706952" watchObservedRunningTime="2025-11-24 14:36:34.686121951 +0000 UTC m=+1031.802762418" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.525165 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.568138 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" event={"ID":"64db7204-1d9c-48c9-92b7-1fb1d468aa6c","Type":"ContainerStarted","Data":"30ce026519206fc63bbda3ddea7bd12a39e42314a17b62fc5b58b00fedf4d15c"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.569405 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.575437 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-n9zjj" event={"ID":"e6c5c30b-4d09-4d55-b483-a6b1a6cfee86","Type":"ContainerStarted","Data":"89d971ff188f31a59bedb48ce2c823221f3141c413ef831842984ab2216d1fad"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.587187 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"8f57fdf8-920a-4329-a0b1-89b62cfa82a6","Type":"ContainerStarted","Data":"95b51b9333a36c5ebae6da3d3fb8108f92bea2d5c2630cd39a2908c7392e032f"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.601277 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" podStartSLOduration=12.146179886 podStartE2EDuration="21.601260836s" podCreationTimestamp="2025-11-24 14:36:14 +0000 UTC" firstStartedPulling="2025-11-24 14:36:24.447267606 +0000 UTC m=+1021.563908083" lastFinishedPulling="2025-11-24 14:36:33.902348556 +0000 UTC m=+1031.018989033" observedRunningTime="2025-11-24 14:36:35.595498339 +0000 UTC m=+1032.712138856" watchObservedRunningTime="2025-11-24 14:36:35.601260836 +0000 UTC m=+1032.717901323" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.622708 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.648663 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7l6s7" event={"ID":"fc49c420-99e9-498c-8b80-c7d2b2c9ed18","Type":"ContainerStarted","Data":"550c762c2dbcdb63c4036b1a5e683402082ff02ff91a36f6987476e60deb8d86"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.648966 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.649129 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.653798 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"ac81e91d-a816-4a86-98b6-3e43f36c7b6f","Type":"ContainerStarted","Data":"b55e78f95a0a85f42d1ff9bbbeb17498fd9a0e18ad8f051174b9962568348cbc"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.654782 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.655935 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" event={"ID":"f03659dc-ab31-4c65-90d4-a94c4044b76e","Type":"ContainerStarted","Data":"a8fbac13800d98099c46c692f9fa50bcf220b100d7d88f9585e06a2a56ecd165"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.656506 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.669242 4822 generic.go:334] "Generic (PLEG): container finished" podID="389f0d9e-a5ab-4d8b-82e2-f4c90d830c42" containerID="5fca1ba42e03886e600ab4c599c29b43c6ef9b26c750f6df2bb1976d457df699" exitCode=0 Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.669339 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42","Type":"ContainerDied","Data":"5fca1ba42e03886e600ab4c599c29b43c6ef9b26c750f6df2bb1976d457df699"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.686729 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-n9zjj" podStartSLOduration=6.509987157 podStartE2EDuration="15.686708905s" podCreationTimestamp="2025-11-24 14:36:20 +0000 UTC" firstStartedPulling="2025-11-24 14:36:24.608758314 +0000 UTC m=+1021.725398801" lastFinishedPulling="2025-11-24 14:36:33.785480072 +0000 UTC m=+1030.902120549" observedRunningTime="2025-11-24 14:36:35.669492985 +0000 UTC m=+1032.786133472" watchObservedRunningTime="2025-11-24 14:36:35.686708905 +0000 UTC m=+1032.803349382" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.699149 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" event={"ID":"66287d73-c11d-48ba-aba2-570671d308e2","Type":"ContainerStarted","Data":"ee0e43423792f4a4ee84bb92c8138b0b8506dc2cfbf349d1b6967c7d3066facb"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.700193 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.720103 4822 generic.go:334] "Generic (PLEG): container finished" podID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerID="21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539" exitCode=0 Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.728464 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-index-gateway-0" podStartSLOduration=11.625047874 podStartE2EDuration="20.728439468s" podCreationTimestamp="2025-11-24 14:36:15 +0000 UTC" firstStartedPulling="2025-11-24 14:36:24.61255851 +0000 UTC m=+1021.729198987" lastFinishedPulling="2025-11-24 14:36:33.715950104 +0000 UTC m=+1030.832590581" observedRunningTime="2025-11-24 14:36:35.711850358 +0000 UTC m=+1032.828490835" watchObservedRunningTime="2025-11-24 14:36:35.728439468 +0000 UTC m=+1032.845079945" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.728887 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerDied","Data":"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.745083 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"830f9ad0-3965-4887-b131-df42fdb35db6","Type":"ContainerStarted","Data":"dcdc580a75442a90abc74c1afd625522eaff01e0822c8f02abd55f5df239b227"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.764122 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7ccfg"] Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.770326 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7l6s7" podStartSLOduration=16.3379902 podStartE2EDuration="26.770298285s" podCreationTimestamp="2025-11-24 14:36:09 +0000 UTC" firstStartedPulling="2025-11-24 14:36:13.343879963 +0000 UTC m=+1010.460520440" lastFinishedPulling="2025-11-24 14:36:23.776188048 +0000 UTC m=+1020.892828525" observedRunningTime="2025-11-24 14:36:35.748125423 +0000 UTC m=+1032.864765920" watchObservedRunningTime="2025-11-24 14:36:35.770298285 +0000 UTC m=+1032.886938762" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.782273 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"890c0404-624f-4c2c-a8d8-ca9aad10c7af","Type":"ContainerStarted","Data":"d7915b502d82d691a96db9c1d66ad51a2007a0aae050e0149e0058709942454a"} Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.783967 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" podStartSLOduration=11.339536604 podStartE2EDuration="20.783948015s" podCreationTimestamp="2025-11-24 14:36:15 +0000 UTC" firstStartedPulling="2025-11-24 14:36:24.602441109 +0000 UTC m=+1021.719081586" lastFinishedPulling="2025-11-24 14:36:34.04685253 +0000 UTC m=+1031.163492997" observedRunningTime="2025-11-24 14:36:35.782444428 +0000 UTC m=+1032.899084925" watchObservedRunningTime="2025-11-24 14:36:35.783948015 +0000 UTC m=+1032.900588492" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.833022 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-s9xbh"] Nov 24 14:36:35 crc kubenswrapper[4822]: E1124 14:36:35.833422 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f960c65-fc33-4f09-a525-e0542ce7207c" containerName="init" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.833433 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f960c65-fc33-4f09-a525-e0542ce7207c" containerName="init" Nov 24 14:36:35 crc kubenswrapper[4822]: E1124 14:36:35.833455 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f960c65-fc33-4f09-a525-e0542ce7207c" containerName="dnsmasq-dns" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.833461 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f960c65-fc33-4f09-a525-e0542ce7207c" containerName="dnsmasq-dns" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.833643 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f960c65-fc33-4f09-a525-e0542ce7207c" containerName="dnsmasq-dns" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.834621 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.853468 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-s9xbh"] Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.853988 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" podStartSLOduration=11.42070733 podStartE2EDuration="20.853972018s" podCreationTimestamp="2025-11-24 14:36:15 +0000 UTC" firstStartedPulling="2025-11-24 14:36:24.459625947 +0000 UTC m=+1021.576266434" lastFinishedPulling="2025-11-24 14:36:33.892890645 +0000 UTC m=+1031.009531122" observedRunningTime="2025-11-24 14:36:35.845150077 +0000 UTC m=+1032.961790554" watchObservedRunningTime="2025-11-24 14:36:35.853972018 +0000 UTC m=+1032.970612495" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.870892 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.890888 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-compactor-0" podStartSLOduration=11.397078183 podStartE2EDuration="20.890867233s" podCreationTimestamp="2025-11-24 14:36:15 +0000 UTC" firstStartedPulling="2025-11-24 14:36:24.519646303 +0000 UTC m=+1021.636286790" lastFinishedPulling="2025-11-24 14:36:34.013435363 +0000 UTC m=+1031.130075840" observedRunningTime="2025-11-24 14:36:35.869124705 +0000 UTC m=+1032.985765192" watchObservedRunningTime="2025-11-24 14:36:35.890867233 +0000 UTC m=+1033.007507710" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.903048 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.425854033 podStartE2EDuration="25.903020246s" podCreationTimestamp="2025-11-24 14:36:10 +0000 UTC" firstStartedPulling="2025-11-24 14:36:13.237589434 +0000 UTC m=+1010.354229911" lastFinishedPulling="2025-11-24 14:36:33.714755647 +0000 UTC m=+1030.831396124" observedRunningTime="2025-11-24 14:36:35.89856114 +0000 UTC m=+1033.015201607" watchObservedRunningTime="2025-11-24 14:36:35.903020246 +0000 UTC m=+1033.019660723" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.930574 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.419243241 podStartE2EDuration="24.930559844s" podCreationTimestamp="2025-11-24 14:36:11 +0000 UTC" firstStartedPulling="2025-11-24 14:36:13.52066358 +0000 UTC m=+1010.637304057" lastFinishedPulling="2025-11-24 14:36:34.031980183 +0000 UTC m=+1031.148620660" observedRunningTime="2025-11-24 14:36:35.921701872 +0000 UTC m=+1033.038342359" watchObservedRunningTime="2025-11-24 14:36:35.930559844 +0000 UTC m=+1033.047200321" Nov 24 14:36:35 crc kubenswrapper[4822]: I1124 14:36:35.958719 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.019566 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-dns-svc\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.019626 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.020553 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swh9m\" (UniqueName: \"kubernetes.io/projected/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-kube-api-access-swh9m\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.020576 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.020592 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-config\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.122929 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-dns-svc\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.123006 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.123037 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swh9m\" (UniqueName: \"kubernetes.io/projected/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-kube-api-access-swh9m\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.123057 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.123075 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-config\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.124105 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.124226 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.124134 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-dns-svc\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.124685 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-config\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.145941 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swh9m\" (UniqueName: \"kubernetes.io/projected/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-kube-api-access-swh9m\") pod \"dnsmasq-dns-698758b865-s9xbh\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.167622 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.608830 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-s9xbh"] Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.797371 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-s9xbh" event={"ID":"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176","Type":"ContainerStarted","Data":"7f0bd4c931b4031a77f2db9c356ba6fe0fc79de1c8093280e06ec20018a580d5"} Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.797722 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" podUID="67610eb7-59cf-4ee6-bc07-0645cd3f41d1" containerName="dnsmasq-dns" containerID="cri-o://18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c" gracePeriod=10 Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.800761 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.842571 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.850670 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.886338 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.887244 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.890435 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-d8mh6" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.890713 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.890838 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.925756 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 14:36:36 crc kubenswrapper[4822]: I1124 14:36:36.969566 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.021499 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.062641 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.062698 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fmbn\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-kube-api-access-2fmbn\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.062900 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4acc7e6a-472b-468a-b709-183f8b3c2b5b-cache\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.063199 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4acc7e6a-472b-468a-b709-183f8b3c2b5b-lock\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.063261 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3a35424b-450a-462a-8d79-3908433569c4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a35424b-450a-462a-8d79-3908433569c4\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.165329 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4acc7e6a-472b-468a-b709-183f8b3c2b5b-lock\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.165714 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3a35424b-450a-462a-8d79-3908433569c4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a35424b-450a-462a-8d79-3908433569c4\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.165804 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.165827 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fmbn\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-kube-api-access-2fmbn\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.165870 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4acc7e6a-472b-468a-b709-183f8b3c2b5b-cache\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.166581 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4acc7e6a-472b-468a-b709-183f8b3c2b5b-cache\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: E1124 14:36:37.166703 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:36:37 crc kubenswrapper[4822]: E1124 14:36:37.166717 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:36:37 crc kubenswrapper[4822]: E1124 14:36:37.166755 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:36:37.666740072 +0000 UTC m=+1034.783380549 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.167154 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4acc7e6a-472b-468a-b709-183f8b3c2b5b-lock\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.176730 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.176764 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3a35424b-450a-462a-8d79-3908433569c4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a35424b-450a-462a-8d79-3908433569c4\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5622796fc8a93e9cecf767a9e2c8d9dedb96e5d4dd965a39a2095f46fb3e1a7a/globalmount\"" pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.207184 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fmbn\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-kube-api-access-2fmbn\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.221526 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-zb9cf"] Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.223152 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.225455 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.225646 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.226992 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.236791 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zb9cf"] Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.241804 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-zb9cf"] Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.251750 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3a35424b-450a-462a-8d79-3908433569c4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a35424b-450a-462a-8d79-3908433569c4\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: E1124 14:36:37.359769 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-zt9lv ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-zb9cf" podUID="5d50b788-57a7-4a3e-a22f-392d121a0858" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.363786 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.370355 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-scripts\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.370391 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-ring-data-devices\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.370478 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-combined-ca-bundle\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.370637 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt9lv\" (UniqueName: \"kubernetes.io/projected/5d50b788-57a7-4a3e-a22f-392d121a0858-kube-api-access-zt9lv\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.370729 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d50b788-57a7-4a3e-a22f-392d121a0858-etc-swift\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.370783 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-swiftconf\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.371009 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-dispersionconf\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.472682 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-sb\") pod \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.472988 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-dns-svc\") pod \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.473119 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctjnf\" (UniqueName: \"kubernetes.io/projected/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-kube-api-access-ctjnf\") pod \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.473140 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-config\") pod \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.473168 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-nb\") pod \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\" (UID: \"67610eb7-59cf-4ee6-bc07-0645cd3f41d1\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.473941 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d50b788-57a7-4a3e-a22f-392d121a0858-etc-swift\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.473976 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-swiftconf\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.474048 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-dispersionconf\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.474098 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-scripts\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.474116 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-ring-data-devices\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.474139 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-combined-ca-bundle\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.474181 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt9lv\" (UniqueName: \"kubernetes.io/projected/5d50b788-57a7-4a3e-a22f-392d121a0858-kube-api-access-zt9lv\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.474757 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d50b788-57a7-4a3e-a22f-392d121a0858-etc-swift\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.475140 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-ring-data-devices\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.475229 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-scripts\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.487032 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-kube-api-access-ctjnf" (OuterVolumeSpecName: "kube-api-access-ctjnf") pod "67610eb7-59cf-4ee6-bc07-0645cd3f41d1" (UID: "67610eb7-59cf-4ee6-bc07-0645cd3f41d1"). InnerVolumeSpecName "kube-api-access-ctjnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.487242 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-dispersionconf\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.487338 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-swiftconf\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.487447 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-combined-ca-bundle\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.491687 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt9lv\" (UniqueName: \"kubernetes.io/projected/5d50b788-57a7-4a3e-a22f-392d121a0858-kube-api-access-zt9lv\") pod \"swift-ring-rebalance-zb9cf\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.520410 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-config" (OuterVolumeSpecName: "config") pod "67610eb7-59cf-4ee6-bc07-0645cd3f41d1" (UID: "67610eb7-59cf-4ee6-bc07-0645cd3f41d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.527844 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "67610eb7-59cf-4ee6-bc07-0645cd3f41d1" (UID: "67610eb7-59cf-4ee6-bc07-0645cd3f41d1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.528767 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "67610eb7-59cf-4ee6-bc07-0645cd3f41d1" (UID: "67610eb7-59cf-4ee6-bc07-0645cd3f41d1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.533727 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67610eb7-59cf-4ee6-bc07-0645cd3f41d1" (UID: "67610eb7-59cf-4ee6-bc07-0645cd3f41d1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.576362 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.576406 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.576422 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctjnf\" (UniqueName: \"kubernetes.io/projected/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-kube-api-access-ctjnf\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.576437 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.576448 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67610eb7-59cf-4ee6-bc07-0645cd3f41d1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.677878 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:37 crc kubenswrapper[4822]: E1124 14:36:37.678037 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:36:37 crc kubenswrapper[4822]: E1124 14:36:37.678050 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:36:37 crc kubenswrapper[4822]: E1124 14:36:37.678090 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:36:38.678077638 +0000 UTC m=+1035.794718115 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.842863 4822 generic.go:334] "Generic (PLEG): container finished" podID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" containerID="f97bdf86e5b2c340231c05300811df6caf732375c3943c2abf5b0cf33f8a074c" exitCode=0 Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.843238 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-s9xbh" event={"ID":"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176","Type":"ContainerDied","Data":"f97bdf86e5b2c340231c05300811df6caf732375c3943c2abf5b0cf33f8a074c"} Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.865186 4822 generic.go:334] "Generic (PLEG): container finished" podID="67610eb7-59cf-4ee6-bc07-0645cd3f41d1" containerID="18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c" exitCode=0 Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.865310 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.866064 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.866968 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" event={"ID":"67610eb7-59cf-4ee6-bc07-0645cd3f41d1","Type":"ContainerDied","Data":"18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c"} Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.867005 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7ccfg" event={"ID":"67610eb7-59cf-4ee6-bc07-0645cd3f41d1","Type":"ContainerDied","Data":"700bf04e145b17cd7009a3bc80ef630ace650c083763b935ced7b55341778916"} Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.867027 4822 scope.go:117] "RemoveContainer" containerID="18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.869441 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.918735 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.939294 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7ccfg"] Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.944396 4822 scope.go:117] "RemoveContainer" containerID="7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.945715 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7ccfg"] Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.950195 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.983909 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-dispersionconf\") pod \"5d50b788-57a7-4a3e-a22f-392d121a0858\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.984012 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d50b788-57a7-4a3e-a22f-392d121a0858-etc-swift\") pod \"5d50b788-57a7-4a3e-a22f-392d121a0858\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.984038 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt9lv\" (UniqueName: \"kubernetes.io/projected/5d50b788-57a7-4a3e-a22f-392d121a0858-kube-api-access-zt9lv\") pod \"5d50b788-57a7-4a3e-a22f-392d121a0858\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.984059 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-combined-ca-bundle\") pod \"5d50b788-57a7-4a3e-a22f-392d121a0858\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.984124 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-swiftconf\") pod \"5d50b788-57a7-4a3e-a22f-392d121a0858\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.984187 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-scripts\") pod \"5d50b788-57a7-4a3e-a22f-392d121a0858\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.984227 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-ring-data-devices\") pod \"5d50b788-57a7-4a3e-a22f-392d121a0858\" (UID: \"5d50b788-57a7-4a3e-a22f-392d121a0858\") " Nov 24 14:36:37 crc kubenswrapper[4822]: I1124 14:36:37.986679 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d50b788-57a7-4a3e-a22f-392d121a0858-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5d50b788-57a7-4a3e-a22f-392d121a0858" (UID: "5d50b788-57a7-4a3e-a22f-392d121a0858"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:37.997905 4822 scope.go:117] "RemoveContainer" containerID="18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.000274 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5d50b788-57a7-4a3e-a22f-392d121a0858" (UID: "5d50b788-57a7-4a3e-a22f-392d121a0858"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.004750 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-scripts" (OuterVolumeSpecName: "scripts") pod "5d50b788-57a7-4a3e-a22f-392d121a0858" (UID: "5d50b788-57a7-4a3e-a22f-392d121a0858"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.016815 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d50b788-57a7-4a3e-a22f-392d121a0858-kube-api-access-zt9lv" (OuterVolumeSpecName: "kube-api-access-zt9lv") pod "5d50b788-57a7-4a3e-a22f-392d121a0858" (UID: "5d50b788-57a7-4a3e-a22f-392d121a0858"). InnerVolumeSpecName "kube-api-access-zt9lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:38 crc kubenswrapper[4822]: E1124 14:36:38.017293 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c\": container with ID starting with 18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c not found: ID does not exist" containerID="18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.017327 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c"} err="failed to get container status \"18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c\": rpc error: code = NotFound desc = could not find container \"18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c\": container with ID starting with 18ca8ac7d4b2a4bd03b7ae0142884f206e769f5da62ac6f75a6baa1ebf1ac41c not found: ID does not exist" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.017351 4822 scope.go:117] "RemoveContainer" containerID="7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.017335 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5d50b788-57a7-4a3e-a22f-392d121a0858" (UID: "5d50b788-57a7-4a3e-a22f-392d121a0858"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.018025 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d50b788-57a7-4a3e-a22f-392d121a0858" (UID: "5d50b788-57a7-4a3e-a22f-392d121a0858"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:36:38 crc kubenswrapper[4822]: E1124 14:36:38.018230 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd\": container with ID starting with 7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd not found: ID does not exist" containerID="7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.018255 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd"} err="failed to get container status \"7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd\": rpc error: code = NotFound desc = could not find container \"7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd\": container with ID starting with 7f0370e2926df4138abcfd0a9846f504c09b5e41124452a4663f7161e5650ebd not found: ID does not exist" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.019322 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5d50b788-57a7-4a3e-a22f-392d121a0858" (UID: "5d50b788-57a7-4a3e-a22f-392d121a0858"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.086592 4822 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.086619 4822 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d50b788-57a7-4a3e-a22f-392d121a0858-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.086629 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt9lv\" (UniqueName: \"kubernetes.io/projected/5d50b788-57a7-4a3e-a22f-392d121a0858-kube-api-access-zt9lv\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.086638 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.086649 4822 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d50b788-57a7-4a3e-a22f-392d121a0858-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.086658 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.086667 4822 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d50b788-57a7-4a3e-a22f-392d121a0858-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.263410 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 14:36:38 crc kubenswrapper[4822]: E1124 14:36:38.263776 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67610eb7-59cf-4ee6-bc07-0645cd3f41d1" containerName="init" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.263788 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="67610eb7-59cf-4ee6-bc07-0645cd3f41d1" containerName="init" Nov 24 14:36:38 crc kubenswrapper[4822]: E1124 14:36:38.263810 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67610eb7-59cf-4ee6-bc07-0645cd3f41d1" containerName="dnsmasq-dns" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.263884 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="67610eb7-59cf-4ee6-bc07-0645cd3f41d1" containerName="dnsmasq-dns" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.264037 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="67610eb7-59cf-4ee6-bc07-0645cd3f41d1" containerName="dnsmasq-dns" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.265020 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.268916 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.269123 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-nt5tw" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.269257 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.269394 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.293071 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.394922 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmwl6\" (UniqueName: \"kubernetes.io/projected/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-kube-api-access-lmwl6\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.394993 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.395145 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.395307 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-config\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.395412 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.395446 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.395467 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-scripts\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.498835 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-config\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.498994 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-config\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.500484 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.500548 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.500583 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-scripts\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.500642 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmwl6\" (UniqueName: \"kubernetes.io/projected/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-kube-api-access-lmwl6\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.500700 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.500778 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.502484 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-scripts\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.502862 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.511000 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.511697 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.511967 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.521346 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmwl6\" (UniqueName: \"kubernetes.io/projected/a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a-kube-api-access-lmwl6\") pod \"ovn-northd-0\" (UID: \"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a\") " pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.590937 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.704459 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:38 crc kubenswrapper[4822]: E1124 14:36:38.704663 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:36:38 crc kubenswrapper[4822]: E1124 14:36:38.704676 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:36:38 crc kubenswrapper[4822]: E1124 14:36:38.704717 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:36:40.704703932 +0000 UTC m=+1037.821344409 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.905060 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-s9xbh" event={"ID":"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176","Type":"ContainerStarted","Data":"5df2a9dad76627891e3173a6132be88f4f251a88fcb642e77f4cfb5d40008de3"} Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.905875 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.916452 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zb9cf" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.927985 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-s9xbh" podStartSLOduration=3.9279689490000003 podStartE2EDuration="3.927968949s" podCreationTimestamp="2025-11-24 14:36:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:36:38.923882313 +0000 UTC m=+1036.040522800" watchObservedRunningTime="2025-11-24 14:36:38.927968949 +0000 UTC m=+1036.044609426" Nov 24 14:36:38 crc kubenswrapper[4822]: I1124 14:36:38.982862 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-zb9cf"] Nov 24 14:36:39 crc kubenswrapper[4822]: I1124 14:36:39.005125 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-zb9cf"] Nov 24 14:36:39 crc kubenswrapper[4822]: E1124 14:36:39.133807 4822 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d50b788_57a7_4a3e_a22f_392d121a0858.slice\": RecentStats: unable to find data in memory cache]" Nov 24 14:36:39 crc kubenswrapper[4822]: I1124 14:36:39.602456 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 14:36:39 crc kubenswrapper[4822]: W1124 14:36:39.613728 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1e7b65a_52b5_4f4b_aa89_f6247e4fa31a.slice/crio-9cee7c7dd2c61374a4507865fcdc062905c23aefcdf48629daa686350175d05c WatchSource:0}: Error finding container 9cee7c7dd2c61374a4507865fcdc062905c23aefcdf48629daa686350175d05c: Status 404 returned error can't find the container with id 9cee7c7dd2c61374a4507865fcdc062905c23aefcdf48629daa686350175d05c Nov 24 14:36:39 crc kubenswrapper[4822]: I1124 14:36:39.720776 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d50b788-57a7-4a3e-a22f-392d121a0858" path="/var/lib/kubelet/pods/5d50b788-57a7-4a3e-a22f-392d121a0858/volumes" Nov 24 14:36:39 crc kubenswrapper[4822]: I1124 14:36:39.721721 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67610eb7-59cf-4ee6-bc07-0645cd3f41d1" path="/var/lib/kubelet/pods/67610eb7-59cf-4ee6-bc07-0645cd3f41d1/volumes" Nov 24 14:36:39 crc kubenswrapper[4822]: I1124 14:36:39.944817 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a","Type":"ContainerStarted","Data":"9cee7c7dd2c61374a4507865fcdc062905c23aefcdf48629daa686350175d05c"} Nov 24 14:36:39 crc kubenswrapper[4822]: I1124 14:36:39.947941 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42","Type":"ContainerStarted","Data":"a0bba7abc45cf8cc5c2dfe783bc35be02fa800120aca030f7788d6778b64d25d"} Nov 24 14:36:40 crc kubenswrapper[4822]: I1124 14:36:40.749668 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:40 crc kubenswrapper[4822]: E1124 14:36:40.749873 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:36:40 crc kubenswrapper[4822]: E1124 14:36:40.749896 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:36:40 crc kubenswrapper[4822]: E1124 14:36:40.749943 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:36:44.749928604 +0000 UTC m=+1041.866569081 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:36:42 crc kubenswrapper[4822]: I1124 14:36:42.406471 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 14:36:42 crc kubenswrapper[4822]: I1124 14:36:42.406819 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 14:36:42 crc kubenswrapper[4822]: I1124 14:36:42.485985 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 14:36:42 crc kubenswrapper[4822]: I1124 14:36:42.988303 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a","Type":"ContainerStarted","Data":"f820d77dc0a36a6bd4df803f3a89db4bcbdcc3f7d438a2407a589bcf9ab6b1fc"} Nov 24 14:36:42 crc kubenswrapper[4822]: I1124 14:36:42.992290 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"389f0d9e-a5ab-4d8b-82e2-f4c90d830c42","Type":"ContainerStarted","Data":"41838a24bebbca17e982895af1969c3622a1d986f651a7a941e6728f5b9001f3"} Nov 24 14:36:42 crc kubenswrapper[4822]: I1124 14:36:42.993415 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.006331 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.008639 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerStarted","Data":"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954"} Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.017030 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=11.75853958 podStartE2EDuration="38.017016739s" podCreationTimestamp="2025-11-24 14:36:05 +0000 UTC" firstStartedPulling="2025-11-24 14:36:13.106891805 +0000 UTC m=+1010.223532282" lastFinishedPulling="2025-11-24 14:36:39.365368964 +0000 UTC m=+1036.482009441" observedRunningTime="2025-11-24 14:36:43.01330568 +0000 UTC m=+1040.129946167" watchObservedRunningTime="2025-11-24 14:36:43.017016739 +0000 UTC m=+1040.133657216" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.149672 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.329601 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.329904 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.407697 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.607661 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-qwctq"] Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.609188 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qwctq" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.629497 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7615-account-create-wzxch"] Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.631075 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.634103 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.655594 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-qwctq"] Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.678139 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7615-account-create-wzxch"] Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.724494 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l44vg\" (UniqueName: \"kubernetes.io/projected/103881b0-811b-4c80-9ca4-d6a105da608f-kube-api-access-l44vg\") pod \"placement-db-create-qwctq\" (UID: \"103881b0-811b-4c80-9ca4-d6a105da608f\") " pod="openstack/placement-db-create-qwctq" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.724561 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103881b0-811b-4c80-9ca4-d6a105da608f-operator-scripts\") pod \"placement-db-create-qwctq\" (UID: \"103881b0-811b-4c80-9ca4-d6a105da608f\") " pod="openstack/placement-db-create-qwctq" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.724687 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnw94\" (UniqueName: \"kubernetes.io/projected/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-kube-api-access-gnw94\") pod \"placement-7615-account-create-wzxch\" (UID: \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\") " pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.724742 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-operator-scripts\") pod \"placement-7615-account-create-wzxch\" (UID: \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\") " pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.787054 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-tmfgk"] Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.788360 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.799217 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-tmfgk"] Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.826376 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-operator-scripts\") pod \"placement-7615-account-create-wzxch\" (UID: \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\") " pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.826749 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l44vg\" (UniqueName: \"kubernetes.io/projected/103881b0-811b-4c80-9ca4-d6a105da608f-kube-api-access-l44vg\") pod \"placement-db-create-qwctq\" (UID: \"103881b0-811b-4c80-9ca4-d6a105da608f\") " pod="openstack/placement-db-create-qwctq" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.826844 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103881b0-811b-4c80-9ca4-d6a105da608f-operator-scripts\") pod \"placement-db-create-qwctq\" (UID: \"103881b0-811b-4c80-9ca4-d6a105da608f\") " pod="openstack/placement-db-create-qwctq" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.827142 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnw94\" (UniqueName: \"kubernetes.io/projected/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-kube-api-access-gnw94\") pod \"placement-7615-account-create-wzxch\" (UID: \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\") " pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.827235 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-operator-scripts\") pod \"placement-7615-account-create-wzxch\" (UID: \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\") " pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.827779 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103881b0-811b-4c80-9ca4-d6a105da608f-operator-scripts\") pod \"placement-db-create-qwctq\" (UID: \"103881b0-811b-4c80-9ca4-d6a105da608f\") " pod="openstack/placement-db-create-qwctq" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.845891 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l44vg\" (UniqueName: \"kubernetes.io/projected/103881b0-811b-4c80-9ca4-d6a105da608f-kube-api-access-l44vg\") pod \"placement-db-create-qwctq\" (UID: \"103881b0-811b-4c80-9ca4-d6a105da608f\") " pod="openstack/placement-db-create-qwctq" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.849918 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnw94\" (UniqueName: \"kubernetes.io/projected/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-kube-api-access-gnw94\") pod \"placement-7615-account-create-wzxch\" (UID: \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\") " pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.899165 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-bfd9-account-create-khpb9"] Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.900555 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.902413 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.909407 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-bfd9-account-create-khpb9"] Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.949647 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn66p\" (UniqueName: \"kubernetes.io/projected/2e3ccadd-2328-41b6-88d8-b833f08073f8-kube-api-access-rn66p\") pod \"glance-db-create-tmfgk\" (UID: \"2e3ccadd-2328-41b6-88d8-b833f08073f8\") " pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.949688 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3ccadd-2328-41b6-88d8-b833f08073f8-operator-scripts\") pod \"glance-db-create-tmfgk\" (UID: \"2e3ccadd-2328-41b6-88d8-b833f08073f8\") " pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:43 crc kubenswrapper[4822]: I1124 14:36:43.996654 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qwctq" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.004056 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.020363 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a","Type":"ContainerStarted","Data":"5695619a1e75f8bd284fbcdf6530c62bf50455ff91544284adb88b53681eae86"} Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.042284 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.05191317 podStartE2EDuration="6.042268403s" podCreationTimestamp="2025-11-24 14:36:38 +0000 UTC" firstStartedPulling="2025-11-24 14:36:39.619249581 +0000 UTC m=+1036.735890058" lastFinishedPulling="2025-11-24 14:36:42.609604774 +0000 UTC m=+1039.726245291" observedRunningTime="2025-11-24 14:36:44.037140413 +0000 UTC m=+1041.153780890" watchObservedRunningTime="2025-11-24 14:36:44.042268403 +0000 UTC m=+1041.158908880" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.051278 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn66p\" (UniqueName: \"kubernetes.io/projected/2e3ccadd-2328-41b6-88d8-b833f08073f8-kube-api-access-rn66p\") pod \"glance-db-create-tmfgk\" (UID: \"2e3ccadd-2328-41b6-88d8-b833f08073f8\") " pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.051316 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3ccadd-2328-41b6-88d8-b833f08073f8-operator-scripts\") pod \"glance-db-create-tmfgk\" (UID: \"2e3ccadd-2328-41b6-88d8-b833f08073f8\") " pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.051354 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df77767f-08e4-4292-94b1-98d55432c038-operator-scripts\") pod \"glance-bfd9-account-create-khpb9\" (UID: \"df77767f-08e4-4292-94b1-98d55432c038\") " pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.051384 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkbtx\" (UniqueName: \"kubernetes.io/projected/df77767f-08e4-4292-94b1-98d55432c038-kube-api-access-pkbtx\") pod \"glance-bfd9-account-create-khpb9\" (UID: \"df77767f-08e4-4292-94b1-98d55432c038\") " pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.052082 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3ccadd-2328-41b6-88d8-b833f08073f8-operator-scripts\") pod \"glance-db-create-tmfgk\" (UID: \"2e3ccadd-2328-41b6-88d8-b833f08073f8\") " pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.073169 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn66p\" (UniqueName: \"kubernetes.io/projected/2e3ccadd-2328-41b6-88d8-b833f08073f8-kube-api-access-rn66p\") pod \"glance-db-create-tmfgk\" (UID: \"2e3ccadd-2328-41b6-88d8-b833f08073f8\") " pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.106295 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.131376 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.156614 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df77767f-08e4-4292-94b1-98d55432c038-operator-scripts\") pod \"glance-bfd9-account-create-khpb9\" (UID: \"df77767f-08e4-4292-94b1-98d55432c038\") " pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.156835 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkbtx\" (UniqueName: \"kubernetes.io/projected/df77767f-08e4-4292-94b1-98d55432c038-kube-api-access-pkbtx\") pod \"glance-bfd9-account-create-khpb9\" (UID: \"df77767f-08e4-4292-94b1-98d55432c038\") " pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.163227 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df77767f-08e4-4292-94b1-98d55432c038-operator-scripts\") pod \"glance-bfd9-account-create-khpb9\" (UID: \"df77767f-08e4-4292-94b1-98d55432c038\") " pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.185971 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkbtx\" (UniqueName: \"kubernetes.io/projected/df77767f-08e4-4292-94b1-98d55432c038-kube-api-access-pkbtx\") pod \"glance-bfd9-account-create-khpb9\" (UID: \"df77767f-08e4-4292-94b1-98d55432c038\") " pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.274901 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.511335 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7615-account-create-wzxch"] Nov 24 14:36:44 crc kubenswrapper[4822]: W1124 14:36:44.513822 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2569eca8_4bd8_4fcb_a9dd_c710ee60c6a6.slice/crio-9e059b92764e22581f0dadedf1a80bc5370079b238daa72485a7e0fc2aa75578 WatchSource:0}: Error finding container 9e059b92764e22581f0dadedf1a80bc5370079b238daa72485a7e0fc2aa75578: Status 404 returned error can't find the container with id 9e059b92764e22581f0dadedf1a80bc5370079b238daa72485a7e0fc2aa75578 Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.525223 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-qwctq"] Nov 24 14:36:44 crc kubenswrapper[4822]: W1124 14:36:44.541979 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod103881b0_811b_4c80_9ca4_d6a105da608f.slice/crio-a418c417dc16fe8b2bfe0ef74a0902ac3321d6791011621542c485b2d03fea05 WatchSource:0}: Error finding container a418c417dc16fe8b2bfe0ef74a0902ac3321d6791011621542c485b2d03fea05: Status 404 returned error can't find the container with id a418c417dc16fe8b2bfe0ef74a0902ac3321d6791011621542c485b2d03fea05 Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.718841 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-bfd9-account-create-khpb9"] Nov 24 14:36:44 crc kubenswrapper[4822]: W1124 14:36:44.725572 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e3ccadd_2328_41b6_88d8_b833f08073f8.slice/crio-c18b1d9e9331e766e46367827228c2b629bcaa95e5e9a6b6234fc8a2dffc60f8 WatchSource:0}: Error finding container c18b1d9e9331e766e46367827228c2b629bcaa95e5e9a6b6234fc8a2dffc60f8: Status 404 returned error can't find the container with id c18b1d9e9331e766e46367827228c2b629bcaa95e5e9a6b6234fc8a2dffc60f8 Nov 24 14:36:44 crc kubenswrapper[4822]: W1124 14:36:44.730786 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf77767f_08e4_4292_94b1_98d55432c038.slice/crio-afda2f95cd4c21f94b2a0d5393c362a12edf4d0e25aba0e7b5d317fb2e3bfcd8 WatchSource:0}: Error finding container afda2f95cd4c21f94b2a0d5393c362a12edf4d0e25aba0e7b5d317fb2e3bfcd8: Status 404 returned error can't find the container with id afda2f95cd4c21f94b2a0d5393c362a12edf4d0e25aba0e7b5d317fb2e3bfcd8 Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.731619 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-tmfgk"] Nov 24 14:36:44 crc kubenswrapper[4822]: I1124 14:36:44.770898 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:44 crc kubenswrapper[4822]: E1124 14:36:44.771131 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:36:44 crc kubenswrapper[4822]: E1124 14:36:44.771167 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:36:44 crc kubenswrapper[4822]: E1124 14:36:44.771252 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:36:52.771230809 +0000 UTC m=+1049.887871296 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.032453 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qwctq" event={"ID":"103881b0-811b-4c80-9ca4-d6a105da608f","Type":"ContainerStarted","Data":"632197d2fbb1062cdb8cead4c3f4fda86897711b08e9c85541fcfcca3e1c3d09"} Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.032512 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qwctq" event={"ID":"103881b0-811b-4c80-9ca4-d6a105da608f","Type":"ContainerStarted","Data":"a418c417dc16fe8b2bfe0ef74a0902ac3321d6791011621542c485b2d03fea05"} Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.034837 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bfd9-account-create-khpb9" event={"ID":"df77767f-08e4-4292-94b1-98d55432c038","Type":"ContainerStarted","Data":"8eae7a022b1425e37963866e5fb38f9f9138c0366bfe83c9fa3e25e457ae6120"} Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.034906 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bfd9-account-create-khpb9" event={"ID":"df77767f-08e4-4292-94b1-98d55432c038","Type":"ContainerStarted","Data":"afda2f95cd4c21f94b2a0d5393c362a12edf4d0e25aba0e7b5d317fb2e3bfcd8"} Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.037737 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7615-account-create-wzxch" event={"ID":"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6","Type":"ContainerStarted","Data":"1d48777a560e08f1d631ef9c9bb5c89fa5c4a6ebce7b125a6b75ae8e7249b12d"} Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.037806 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7615-account-create-wzxch" event={"ID":"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6","Type":"ContainerStarted","Data":"9e059b92764e22581f0dadedf1a80bc5370079b238daa72485a7e0fc2aa75578"} Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.039713 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tmfgk" event={"ID":"2e3ccadd-2328-41b6-88d8-b833f08073f8","Type":"ContainerStarted","Data":"046279f7bbddb28e4e582ee2dde1a05edf5e7e1d995cab69ba63764c65fbde1c"} Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.039759 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tmfgk" event={"ID":"2e3ccadd-2328-41b6-88d8-b833f08073f8","Type":"ContainerStarted","Data":"c18b1d9e9331e766e46367827228c2b629bcaa95e5e9a6b6234fc8a2dffc60f8"} Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.040355 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.048255 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-qwctq" podStartSLOduration=2.048235394 podStartE2EDuration="2.048235394s" podCreationTimestamp="2025-11-24 14:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:36:45.045272227 +0000 UTC m=+1042.161912714" watchObservedRunningTime="2025-11-24 14:36:45.048235394 +0000 UTC m=+1042.164875871" Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.067514 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7615-account-create-wzxch" podStartSLOduration=2.067496028 podStartE2EDuration="2.067496028s" podCreationTimestamp="2025-11-24 14:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:36:45.063365346 +0000 UTC m=+1042.180005863" watchObservedRunningTime="2025-11-24 14:36:45.067496028 +0000 UTC m=+1042.184136505" Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.086959 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-bfd9-account-create-khpb9" podStartSLOduration=2.086939537 podStartE2EDuration="2.086939537s" podCreationTimestamp="2025-11-24 14:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:36:45.07749624 +0000 UTC m=+1042.194136727" watchObservedRunningTime="2025-11-24 14:36:45.086939537 +0000 UTC m=+1042.203580024" Nov 24 14:36:45 crc kubenswrapper[4822]: I1124 14:36:45.105822 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-tmfgk" podStartSLOduration=2.10580442 podStartE2EDuration="2.10580442s" podCreationTimestamp="2025-11-24 14:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:36:45.097360462 +0000 UTC m=+1042.214000949" watchObservedRunningTime="2025-11-24 14:36:45.10580442 +0000 UTC m=+1042.222444907" Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.059717 4822 generic.go:334] "Generic (PLEG): container finished" podID="2e3ccadd-2328-41b6-88d8-b833f08073f8" containerID="046279f7bbddb28e4e582ee2dde1a05edf5e7e1d995cab69ba63764c65fbde1c" exitCode=0 Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.059836 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tmfgk" event={"ID":"2e3ccadd-2328-41b6-88d8-b833f08073f8","Type":"ContainerDied","Data":"046279f7bbddb28e4e582ee2dde1a05edf5e7e1d995cab69ba63764c65fbde1c"} Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.063711 4822 generic.go:334] "Generic (PLEG): container finished" podID="103881b0-811b-4c80-9ca4-d6a105da608f" containerID="632197d2fbb1062cdb8cead4c3f4fda86897711b08e9c85541fcfcca3e1c3d09" exitCode=0 Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.063751 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qwctq" event={"ID":"103881b0-811b-4c80-9ca4-d6a105da608f","Type":"ContainerDied","Data":"632197d2fbb1062cdb8cead4c3f4fda86897711b08e9c85541fcfcca3e1c3d09"} Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.066142 4822 generic.go:334] "Generic (PLEG): container finished" podID="df77767f-08e4-4292-94b1-98d55432c038" containerID="8eae7a022b1425e37963866e5fb38f9f9138c0366bfe83c9fa3e25e457ae6120" exitCode=0 Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.066237 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bfd9-account-create-khpb9" event={"ID":"df77767f-08e4-4292-94b1-98d55432c038","Type":"ContainerDied","Data":"8eae7a022b1425e37963866e5fb38f9f9138c0366bfe83c9fa3e25e457ae6120"} Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.073618 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerStarted","Data":"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9"} Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.093388 4822 generic.go:334] "Generic (PLEG): container finished" podID="2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6" containerID="1d48777a560e08f1d631ef9c9bb5c89fa5c4a6ebce7b125a6b75ae8e7249b12d" exitCode=0 Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.093518 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7615-account-create-wzxch" event={"ID":"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6","Type":"ContainerDied","Data":"1d48777a560e08f1d631ef9c9bb5c89fa5c4a6ebce7b125a6b75ae8e7249b12d"} Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.169400 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.249246 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfn5c"] Nov 24 14:36:46 crc kubenswrapper[4822]: I1124 14:36:46.253691 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" podUID="a16b9b73-6577-4420-89bb-3c774f9c4bd3" containerName="dnsmasq-dns" containerID="cri-o://61f573942a5401933034ae3d1ba91c22291a15242f22a603ea0f4d737cacdb03" gracePeriod=10 Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.102702 4822 generic.go:334] "Generic (PLEG): container finished" podID="a16b9b73-6577-4420-89bb-3c774f9c4bd3" containerID="61f573942a5401933034ae3d1ba91c22291a15242f22a603ea0f4d737cacdb03" exitCode=0 Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.102980 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" event={"ID":"a16b9b73-6577-4420-89bb-3c774f9c4bd3","Type":"ContainerDied","Data":"61f573942a5401933034ae3d1ba91c22291a15242f22a603ea0f4d737cacdb03"} Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.369702 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.428990 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-config\") pod \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.429088 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nj6g\" (UniqueName: \"kubernetes.io/projected/a16b9b73-6577-4420-89bb-3c774f9c4bd3-kube-api-access-2nj6g\") pod \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.429166 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-dns-svc\") pod \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\" (UID: \"a16b9b73-6577-4420-89bb-3c774f9c4bd3\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.451495 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a16b9b73-6577-4420-89bb-3c774f9c4bd3-kube-api-access-2nj6g" (OuterVolumeSpecName: "kube-api-access-2nj6g") pod "a16b9b73-6577-4420-89bb-3c774f9c4bd3" (UID: "a16b9b73-6577-4420-89bb-3c774f9c4bd3"). InnerVolumeSpecName "kube-api-access-2nj6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.478192 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-config" (OuterVolumeSpecName: "config") pod "a16b9b73-6577-4420-89bb-3c774f9c4bd3" (UID: "a16b9b73-6577-4420-89bb-3c774f9c4bd3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.503062 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a16b9b73-6577-4420-89bb-3c774f9c4bd3" (UID: "a16b9b73-6577-4420-89bb-3c774f9c4bd3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.533936 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.533971 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nj6g\" (UniqueName: \"kubernetes.io/projected/a16b9b73-6577-4420-89bb-3c774f9c4bd3-kube-api-access-2nj6g\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.533986 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a16b9b73-6577-4420-89bb-3c774f9c4bd3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.586331 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qwctq" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.635262 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103881b0-811b-4c80-9ca4-d6a105da608f-operator-scripts\") pod \"103881b0-811b-4c80-9ca4-d6a105da608f\" (UID: \"103881b0-811b-4c80-9ca4-d6a105da608f\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.635394 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l44vg\" (UniqueName: \"kubernetes.io/projected/103881b0-811b-4c80-9ca4-d6a105da608f-kube-api-access-l44vg\") pod \"103881b0-811b-4c80-9ca4-d6a105da608f\" (UID: \"103881b0-811b-4c80-9ca4-d6a105da608f\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.636628 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/103881b0-811b-4c80-9ca4-d6a105da608f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "103881b0-811b-4c80-9ca4-d6a105da608f" (UID: "103881b0-811b-4c80-9ca4-d6a105da608f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.662180 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/103881b0-811b-4c80-9ca4-d6a105da608f-kube-api-access-l44vg" (OuterVolumeSpecName: "kube-api-access-l44vg") pod "103881b0-811b-4c80-9ca4-d6a105da608f" (UID: "103881b0-811b-4c80-9ca4-d6a105da608f"). InnerVolumeSpecName "kube-api-access-l44vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.726927 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.730929 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.740674 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103881b0-811b-4c80-9ca4-d6a105da608f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.740697 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l44vg\" (UniqueName: \"kubernetes.io/projected/103881b0-811b-4c80-9ca4-d6a105da608f-kube-api-access-l44vg\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.748103 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.842422 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkbtx\" (UniqueName: \"kubernetes.io/projected/df77767f-08e4-4292-94b1-98d55432c038-kube-api-access-pkbtx\") pod \"df77767f-08e4-4292-94b1-98d55432c038\" (UID: \"df77767f-08e4-4292-94b1-98d55432c038\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.842538 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-operator-scripts\") pod \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\" (UID: \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.842637 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnw94\" (UniqueName: \"kubernetes.io/projected/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-kube-api-access-gnw94\") pod \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\" (UID: \"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.842729 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3ccadd-2328-41b6-88d8-b833f08073f8-operator-scripts\") pod \"2e3ccadd-2328-41b6-88d8-b833f08073f8\" (UID: \"2e3ccadd-2328-41b6-88d8-b833f08073f8\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.842815 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df77767f-08e4-4292-94b1-98d55432c038-operator-scripts\") pod \"df77767f-08e4-4292-94b1-98d55432c038\" (UID: \"df77767f-08e4-4292-94b1-98d55432c038\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.842883 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn66p\" (UniqueName: \"kubernetes.io/projected/2e3ccadd-2328-41b6-88d8-b833f08073f8-kube-api-access-rn66p\") pod \"2e3ccadd-2328-41b6-88d8-b833f08073f8\" (UID: \"2e3ccadd-2328-41b6-88d8-b833f08073f8\") " Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.843323 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6" (UID: "2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.843461 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.843628 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df77767f-08e4-4292-94b1-98d55432c038-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "df77767f-08e4-4292-94b1-98d55432c038" (UID: "df77767f-08e4-4292-94b1-98d55432c038"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.843635 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e3ccadd-2328-41b6-88d8-b833f08073f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e3ccadd-2328-41b6-88d8-b833f08073f8" (UID: "2e3ccadd-2328-41b6-88d8-b833f08073f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.845485 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-kube-api-access-gnw94" (OuterVolumeSpecName: "kube-api-access-gnw94") pod "2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6" (UID: "2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6"). InnerVolumeSpecName "kube-api-access-gnw94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.846508 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df77767f-08e4-4292-94b1-98d55432c038-kube-api-access-pkbtx" (OuterVolumeSpecName: "kube-api-access-pkbtx") pod "df77767f-08e4-4292-94b1-98d55432c038" (UID: "df77767f-08e4-4292-94b1-98d55432c038"). InnerVolumeSpecName "kube-api-access-pkbtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.846935 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e3ccadd-2328-41b6-88d8-b833f08073f8-kube-api-access-rn66p" (OuterVolumeSpecName: "kube-api-access-rn66p") pod "2e3ccadd-2328-41b6-88d8-b833f08073f8" (UID: "2e3ccadd-2328-41b6-88d8-b833f08073f8"). InnerVolumeSpecName "kube-api-access-rn66p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.944363 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkbtx\" (UniqueName: \"kubernetes.io/projected/df77767f-08e4-4292-94b1-98d55432c038-kube-api-access-pkbtx\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.944395 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnw94\" (UniqueName: \"kubernetes.io/projected/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6-kube-api-access-gnw94\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.944407 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3ccadd-2328-41b6-88d8-b833f08073f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.944416 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df77767f-08e4-4292-94b1-98d55432c038-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:47 crc kubenswrapper[4822]: I1124 14:36:47.944426 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn66p\" (UniqueName: \"kubernetes.io/projected/2e3ccadd-2328-41b6-88d8-b833f08073f8-kube-api-access-rn66p\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.122844 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7615-account-create-wzxch" event={"ID":"2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6","Type":"ContainerDied","Data":"9e059b92764e22581f0dadedf1a80bc5370079b238daa72485a7e0fc2aa75578"} Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.122885 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e059b92764e22581f0dadedf1a80bc5370079b238daa72485a7e0fc2aa75578" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.122967 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7615-account-create-wzxch" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.129598 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tmfgk" event={"ID":"2e3ccadd-2328-41b6-88d8-b833f08073f8","Type":"ContainerDied","Data":"c18b1d9e9331e766e46367827228c2b629bcaa95e5e9a6b6234fc8a2dffc60f8"} Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.129639 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c18b1d9e9331e766e46367827228c2b629bcaa95e5e9a6b6234fc8a2dffc60f8" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.129711 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tmfgk" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.134013 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" event={"ID":"a16b9b73-6577-4420-89bb-3c774f9c4bd3","Type":"ContainerDied","Data":"64f90f7e41583677d0c253bdfba33733e9fef891ccd4c174d72419db6baad9ce"} Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.134071 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cfn5c" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.134075 4822 scope.go:117] "RemoveContainer" containerID="61f573942a5401933034ae3d1ba91c22291a15242f22a603ea0f4d737cacdb03" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.140450 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qwctq" event={"ID":"103881b0-811b-4c80-9ca4-d6a105da608f","Type":"ContainerDied","Data":"a418c417dc16fe8b2bfe0ef74a0902ac3321d6791011621542c485b2d03fea05"} Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.140647 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a418c417dc16fe8b2bfe0ef74a0902ac3321d6791011621542c485b2d03fea05" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.140577 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qwctq" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.143016 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bfd9-account-create-khpb9" event={"ID":"df77767f-08e4-4292-94b1-98d55432c038","Type":"ContainerDied","Data":"afda2f95cd4c21f94b2a0d5393c362a12edf4d0e25aba0e7b5d317fb2e3bfcd8"} Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.143065 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afda2f95cd4c21f94b2a0d5393c362a12edf4d0e25aba0e7b5d317fb2e3bfcd8" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.143120 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bfd9-account-create-khpb9" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.169290 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfn5c"] Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.173575 4822 scope.go:117] "RemoveContainer" containerID="86dc9fbc9b39772eca9e0f1d7131deb8ee991c4d5e2fb5f2025c3b6b42025e36" Nov 24 14:36:48 crc kubenswrapper[4822]: I1124 14:36:48.179198 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cfn5c"] Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.064448 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-bwmcx"] Nov 24 14:36:49 crc kubenswrapper[4822]: E1124 14:36:49.065125 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6" containerName="mariadb-account-create" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065153 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6" containerName="mariadb-account-create" Nov 24 14:36:49 crc kubenswrapper[4822]: E1124 14:36:49.065181 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3ccadd-2328-41b6-88d8-b833f08073f8" containerName="mariadb-database-create" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065188 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3ccadd-2328-41b6-88d8-b833f08073f8" containerName="mariadb-database-create" Nov 24 14:36:49 crc kubenswrapper[4822]: E1124 14:36:49.065216 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df77767f-08e4-4292-94b1-98d55432c038" containerName="mariadb-account-create" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065224 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="df77767f-08e4-4292-94b1-98d55432c038" containerName="mariadb-account-create" Nov 24 14:36:49 crc kubenswrapper[4822]: E1124 14:36:49.065233 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16b9b73-6577-4420-89bb-3c774f9c4bd3" containerName="dnsmasq-dns" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065240 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16b9b73-6577-4420-89bb-3c774f9c4bd3" containerName="dnsmasq-dns" Nov 24 14:36:49 crc kubenswrapper[4822]: E1124 14:36:49.065252 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="103881b0-811b-4c80-9ca4-d6a105da608f" containerName="mariadb-database-create" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065258 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="103881b0-811b-4c80-9ca4-d6a105da608f" containerName="mariadb-database-create" Nov 24 14:36:49 crc kubenswrapper[4822]: E1124 14:36:49.065274 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16b9b73-6577-4420-89bb-3c774f9c4bd3" containerName="init" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065282 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16b9b73-6577-4420-89bb-3c774f9c4bd3" containerName="init" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065464 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="df77767f-08e4-4292-94b1-98d55432c038" containerName="mariadb-account-create" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065480 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6" containerName="mariadb-account-create" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065494 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a16b9b73-6577-4420-89bb-3c774f9c4bd3" containerName="dnsmasq-dns" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065511 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3ccadd-2328-41b6-88d8-b833f08073f8" containerName="mariadb-database-create" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.065519 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="103881b0-811b-4c80-9ca4-d6a105da608f" containerName="mariadb-database-create" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.066253 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.070896 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-l2kgh" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.070974 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.078150 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-bwmcx"] Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.166602 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjsvq\" (UniqueName: \"kubernetes.io/projected/a39cbd81-d3b4-46a5-a513-031aca8e43c3-kube-api-access-sjsvq\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.166742 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-db-sync-config-data\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.166967 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-config-data\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.167045 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-combined-ca-bundle\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.267659 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-db-sync-config-data\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.267721 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-config-data\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.267747 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-combined-ca-bundle\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.267843 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjsvq\" (UniqueName: \"kubernetes.io/projected/a39cbd81-d3b4-46a5-a513-031aca8e43c3-kube-api-access-sjsvq\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.272287 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-db-sync-config-data\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.272353 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-combined-ca-bundle\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.273439 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-config-data\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.290559 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjsvq\" (UniqueName: \"kubernetes.io/projected/a39cbd81-d3b4-46a5-a513-031aca8e43c3-kube-api-access-sjsvq\") pod \"glance-db-sync-bwmcx\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.400376 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bwmcx" Nov 24 14:36:49 crc kubenswrapper[4822]: I1124 14:36:49.714699 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a16b9b73-6577-4420-89bb-3c774f9c4bd3" path="/var/lib/kubelet/pods/a16b9b73-6577-4420-89bb-3c774f9c4bd3/volumes" Nov 24 14:36:50 crc kubenswrapper[4822]: I1124 14:36:50.167827 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerStarted","Data":"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c"} Nov 24 14:36:50 crc kubenswrapper[4822]: I1124 14:36:50.199440 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=8.56049997 podStartE2EDuration="45.199419237s" podCreationTimestamp="2025-11-24 14:36:05 +0000 UTC" firstStartedPulling="2025-11-24 14:36:13.11812294 +0000 UTC m=+1010.234763417" lastFinishedPulling="2025-11-24 14:36:49.757042207 +0000 UTC m=+1046.873682684" observedRunningTime="2025-11-24 14:36:50.191953107 +0000 UTC m=+1047.308593624" watchObservedRunningTime="2025-11-24 14:36:50.199419237 +0000 UTC m=+1047.316059724" Nov 24 14:36:50 crc kubenswrapper[4822]: I1124 14:36:50.243676 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-bwmcx"] Nov 24 14:36:51 crc kubenswrapper[4822]: I1124 14:36:51.184290 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bwmcx" event={"ID":"a39cbd81-d3b4-46a5-a513-031aca8e43c3","Type":"ContainerStarted","Data":"51db94409f902bc660472c31e47f64d6d9f70b908f447bab4b8276c96537b1c3"} Nov 24 14:36:51 crc kubenswrapper[4822]: I1124 14:36:51.849109 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:51 crc kubenswrapper[4822]: I1124 14:36:51.849174 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:51 crc kubenswrapper[4822]: I1124 14:36:51.855435 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:52 crc kubenswrapper[4822]: I1124 14:36:52.199188 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:52 crc kubenswrapper[4822]: I1124 14:36:52.780399 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:36:52 crc kubenswrapper[4822]: E1124 14:36:52.780761 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:36:52 crc kubenswrapper[4822]: E1124 14:36:52.780809 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:36:52 crc kubenswrapper[4822]: E1124 14:36:52.780916 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:37:08.78085244 +0000 UTC m=+1065.897492957 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.212388 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-lt4wx"] Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.215915 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.275179 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lt4wx"] Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.305266 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-operator-scripts\") pod \"keystone-db-create-lt4wx\" (UID: \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\") " pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.305424 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdrvf\" (UniqueName: \"kubernetes.io/projected/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-kube-api-access-vdrvf\") pod \"keystone-db-create-lt4wx\" (UID: \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\") " pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.326578 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-72ec-account-create-hgstv"] Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.328320 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.331993 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.333516 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-72ec-account-create-hgstv"] Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.406897 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-operator-scripts\") pod \"keystone-db-create-lt4wx\" (UID: \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\") " pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.407025 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdrvf\" (UniqueName: \"kubernetes.io/projected/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-kube-api-access-vdrvf\") pod \"keystone-db-create-lt4wx\" (UID: \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\") " pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.408070 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-operator-scripts\") pod \"keystone-db-create-lt4wx\" (UID: \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\") " pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.427578 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdrvf\" (UniqueName: \"kubernetes.io/projected/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-kube-api-access-vdrvf\") pod \"keystone-db-create-lt4wx\" (UID: \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\") " pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.509975 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrllx\" (UniqueName: \"kubernetes.io/projected/53ec312d-2011-41f0-8a13-4268f81bfaa6-kube-api-access-rrllx\") pod \"keystone-72ec-account-create-hgstv\" (UID: \"53ec312d-2011-41f0-8a13-4268f81bfaa6\") " pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.510574 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec312d-2011-41f0-8a13-4268f81bfaa6-operator-scripts\") pod \"keystone-72ec-account-create-hgstv\" (UID: \"53ec312d-2011-41f0-8a13-4268f81bfaa6\") " pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.581154 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.612810 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrllx\" (UniqueName: \"kubernetes.io/projected/53ec312d-2011-41f0-8a13-4268f81bfaa6-kube-api-access-rrllx\") pod \"keystone-72ec-account-create-hgstv\" (UID: \"53ec312d-2011-41f0-8a13-4268f81bfaa6\") " pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.612860 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec312d-2011-41f0-8a13-4268f81bfaa6-operator-scripts\") pod \"keystone-72ec-account-create-hgstv\" (UID: \"53ec312d-2011-41f0-8a13-4268f81bfaa6\") " pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.614773 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec312d-2011-41f0-8a13-4268f81bfaa6-operator-scripts\") pod \"keystone-72ec-account-create-hgstv\" (UID: \"53ec312d-2011-41f0-8a13-4268f81bfaa6\") " pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.636040 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrllx\" (UniqueName: \"kubernetes.io/projected/53ec312d-2011-41f0-8a13-4268f81bfaa6-kube-api-access-rrllx\") pod \"keystone-72ec-account-create-hgstv\" (UID: \"53ec312d-2011-41f0-8a13-4268f81bfaa6\") " pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.653215 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:53 crc kubenswrapper[4822]: I1124 14:36:53.720267 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 14:36:54 crc kubenswrapper[4822]: I1124 14:36:54.171666 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lt4wx"] Nov 24 14:36:54 crc kubenswrapper[4822]: I1124 14:36:54.245235 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-72ec-account-create-hgstv"] Nov 24 14:36:54 crc kubenswrapper[4822]: I1124 14:36:54.281363 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lt4wx" event={"ID":"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67","Type":"ContainerStarted","Data":"1fb051c46c40fe74cf5e577131de00c7c2a18b94fb7484cc84075dc08b91dc06"} Nov 24 14:36:54 crc kubenswrapper[4822]: I1124 14:36:54.732816 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.223194 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-mvpjv" Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.296769 4822 generic.go:334] "Generic (PLEG): container finished" podID="ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67" containerID="89511df05f18c959e4d41f1fab6a60626eb20e9dde051f4901b84a8e8f6d32c3" exitCode=0 Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.296840 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lt4wx" event={"ID":"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67","Type":"ContainerDied","Data":"89511df05f18c959e4d41f1fab6a60626eb20e9dde051f4901b84a8e8f6d32c3"} Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.299458 4822 generic.go:334] "Generic (PLEG): container finished" podID="53ec312d-2011-41f0-8a13-4268f81bfaa6" containerID="9f74a7941b7027868abcbfbf89668d256b3f8a261913d05d9a08810b9b88fedc" exitCode=0 Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.299604 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-72ec-account-create-hgstv" event={"ID":"53ec312d-2011-41f0-8a13-4268f81bfaa6","Type":"ContainerDied","Data":"9f74a7941b7027868abcbfbf89668d256b3f8a261913d05d9a08810b9b88fedc"} Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.299659 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-72ec-account-create-hgstv" event={"ID":"53ec312d-2011-41f0-8a13-4268f81bfaa6","Type":"ContainerStarted","Data":"641bbd9e5ecf4de47eae4eedc8a86a18b4d4b7baf032c2e7a50cec0d2a4ab900"} Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.299711 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="prometheus" containerID="cri-o://fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954" gracePeriod=600 Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.299825 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="config-reloader" containerID="cri-o://e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9" gracePeriod=600 Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.299861 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="thanos-sidecar" containerID="cri-o://9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c" gracePeriod=600 Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.514545 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-querier-548665d79b-fjkbp" Nov 24 14:36:55 crc kubenswrapper[4822]: I1124 14:36:55.525354 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-f557q" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.284920 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.318244 4822 generic.go:334] "Generic (PLEG): container finished" podID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerID="9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c" exitCode=0 Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.318290 4822 generic.go:334] "Generic (PLEG): container finished" podID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerID="e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9" exitCode=0 Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.318304 4822 generic.go:334] "Generic (PLEG): container finished" podID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerID="fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954" exitCode=0 Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.318490 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.319181 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerDied","Data":"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c"} Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.319230 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerDied","Data":"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9"} Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.319244 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerDied","Data":"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954"} Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.319257 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"08f13b2a-d10d-4b51-be76-b4c510b24d00","Type":"ContainerDied","Data":"24f52fd79033d679211dfe7f06ada77f47d35d873445b1e5ad8fed0d10d6e535"} Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.319277 4822 scope.go:117] "RemoveContainer" containerID="9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.378935 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz778\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-kube-api-access-zz778\") pod \"08f13b2a-d10d-4b51-be76-b4c510b24d00\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.379292 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/08f13b2a-d10d-4b51-be76-b4c510b24d00-prometheus-metric-storage-rulefiles-0\") pod \"08f13b2a-d10d-4b51-be76-b4c510b24d00\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.379317 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/08f13b2a-d10d-4b51-be76-b4c510b24d00-config-out\") pod \"08f13b2a-d10d-4b51-be76-b4c510b24d00\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.379381 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-config\") pod \"08f13b2a-d10d-4b51-be76-b4c510b24d00\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.379404 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-thanos-prometheus-http-client-file\") pod \"08f13b2a-d10d-4b51-be76-b4c510b24d00\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.379432 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-tls-assets\") pod \"08f13b2a-d10d-4b51-be76-b4c510b24d00\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.379565 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") pod \"08f13b2a-d10d-4b51-be76-b4c510b24d00\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.379631 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-web-config\") pod \"08f13b2a-d10d-4b51-be76-b4c510b24d00\" (UID: \"08f13b2a-d10d-4b51-be76-b4c510b24d00\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.383149 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08f13b2a-d10d-4b51-be76-b4c510b24d00-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "08f13b2a-d10d-4b51-be76-b4c510b24d00" (UID: "08f13b2a-d10d-4b51-be76-b4c510b24d00"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.385552 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-config" (OuterVolumeSpecName: "config") pod "08f13b2a-d10d-4b51-be76-b4c510b24d00" (UID: "08f13b2a-d10d-4b51-be76-b4c510b24d00"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.386628 4822 scope.go:117] "RemoveContainer" containerID="e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.389090 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-kube-api-access-zz778" (OuterVolumeSpecName: "kube-api-access-zz778") pod "08f13b2a-d10d-4b51-be76-b4c510b24d00" (UID: "08f13b2a-d10d-4b51-be76-b4c510b24d00"). InnerVolumeSpecName "kube-api-access-zz778". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.392830 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="849f7f3c-3605-4e1b-8044-5e063841745e" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.392991 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08f13b2a-d10d-4b51-be76-b4c510b24d00-config-out" (OuterVolumeSpecName: "config-out") pod "08f13b2a-d10d-4b51-be76-b4c510b24d00" (UID: "08f13b2a-d10d-4b51-be76-b4c510b24d00"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.394308 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "08f13b2a-d10d-4b51-be76-b4c510b24d00" (UID: "08f13b2a-d10d-4b51-be76-b4c510b24d00"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.394378 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "08f13b2a-d10d-4b51-be76-b4c510b24d00" (UID: "08f13b2a-d10d-4b51-be76-b4c510b24d00"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.410169 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "08f13b2a-d10d-4b51-be76-b4c510b24d00" (UID: "08f13b2a-d10d-4b51-be76-b4c510b24d00"). InnerVolumeSpecName "pvc-c165ae10-7797-4222-877f-f5bb6c853296". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.418391 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-web-config" (OuterVolumeSpecName: "web-config") pod "08f13b2a-d10d-4b51-be76-b4c510b24d00" (UID: "08f13b2a-d10d-4b51-be76-b4c510b24d00"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.465292 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.483409 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz778\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-kube-api-access-zz778\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.483441 4822 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/08f13b2a-d10d-4b51-be76-b4c510b24d00-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.483453 4822 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/08f13b2a-d10d-4b51-be76-b4c510b24d00-config-out\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.483465 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.483474 4822 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.483483 4822 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/08f13b2a-d10d-4b51-be76-b4c510b24d00-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.483543 4822 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") on node \"crc\" " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.483555 4822 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/08f13b2a-d10d-4b51-be76-b4c510b24d00-web-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.501275 4822 scope.go:117] "RemoveContainer" containerID="fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.505440 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.521851 4822 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.522408 4822 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c165ae10-7797-4222-877f-f5bb6c853296" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296") on node "crc" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.551127 4822 scope.go:117] "RemoveContainer" containerID="21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.576781 4822 scope.go:117] "RemoveContainer" containerID="9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c" Nov 24 14:36:56 crc kubenswrapper[4822]: E1124 14:36:56.577303 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c\": container with ID starting with 9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c not found: ID does not exist" containerID="9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.577332 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c"} err="failed to get container status \"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c\": rpc error: code = NotFound desc = could not find container \"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c\": container with ID starting with 9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.577352 4822 scope.go:117] "RemoveContainer" containerID="e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9" Nov 24 14:36:56 crc kubenswrapper[4822]: E1124 14:36:56.577695 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9\": container with ID starting with e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9 not found: ID does not exist" containerID="e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.577723 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9"} err="failed to get container status \"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9\": rpc error: code = NotFound desc = could not find container \"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9\": container with ID starting with e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9 not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.577741 4822 scope.go:117] "RemoveContainer" containerID="fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954" Nov 24 14:36:56 crc kubenswrapper[4822]: E1124 14:36:56.578292 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954\": container with ID starting with fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954 not found: ID does not exist" containerID="fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.578379 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954"} err="failed to get container status \"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954\": rpc error: code = NotFound desc = could not find container \"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954\": container with ID starting with fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954 not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.578394 4822 scope.go:117] "RemoveContainer" containerID="21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539" Nov 24 14:36:56 crc kubenswrapper[4822]: E1124 14:36:56.581500 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539\": container with ID starting with 21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539 not found: ID does not exist" containerID="21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.581540 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539"} err="failed to get container status \"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539\": rpc error: code = NotFound desc = could not find container \"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539\": container with ID starting with 21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539 not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.581572 4822 scope.go:117] "RemoveContainer" containerID="9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.581892 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c"} err="failed to get container status \"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c\": rpc error: code = NotFound desc = could not find container \"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c\": container with ID starting with 9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.581911 4822 scope.go:117] "RemoveContainer" containerID="e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.582928 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9"} err="failed to get container status \"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9\": rpc error: code = NotFound desc = could not find container \"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9\": container with ID starting with e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9 not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.582948 4822 scope.go:117] "RemoveContainer" containerID="fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.583214 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954"} err="failed to get container status \"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954\": rpc error: code = NotFound desc = could not find container \"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954\": container with ID starting with fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954 not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.583231 4822 scope.go:117] "RemoveContainer" containerID="21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.584788 4822 reconciler_common.go:293] "Volume detached for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.585252 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539"} err="failed to get container status \"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539\": rpc error: code = NotFound desc = could not find container \"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539\": container with ID starting with 21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539 not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.585271 4822 scope.go:117] "RemoveContainer" containerID="9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.587726 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c"} err="failed to get container status \"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c\": rpc error: code = NotFound desc = could not find container \"9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c\": container with ID starting with 9d2d8a1d2d191a4922590f512079eaec1f08a4e1c80a5e0c09d98e5c61f9043c not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.587768 4822 scope.go:117] "RemoveContainer" containerID="e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.588267 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9"} err="failed to get container status \"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9\": rpc error: code = NotFound desc = could not find container \"e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9\": container with ID starting with e2700658d2d60601f9399272f859fb1a3c72ad7022087b69bd6673c6d09dc2e9 not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.588284 4822 scope.go:117] "RemoveContainer" containerID="fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.589061 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954"} err="failed to get container status \"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954\": rpc error: code = NotFound desc = could not find container \"fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954\": container with ID starting with fc439771dfb165d672c5821e7c3866fa84519ca4e261831d67dbe91c94440954 not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.589081 4822 scope.go:117] "RemoveContainer" containerID="21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.596338 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539"} err="failed to get container status \"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539\": rpc error: code = NotFound desc = could not find container \"21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539\": container with ID starting with 21061a297a627bed9bdc2cfcdd0584f782d3b67abc0aae3614d9ef06de264539 not found: ID does not exist" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.665517 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.688107 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.701722 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:36:56 crc kubenswrapper[4822]: E1124 14:36:56.702238 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="config-reloader" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.702254 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="config-reloader" Nov 24 14:36:56 crc kubenswrapper[4822]: E1124 14:36:56.702267 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="init-config-reloader" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.702274 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="init-config-reloader" Nov 24 14:36:56 crc kubenswrapper[4822]: E1124 14:36:56.702295 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="thanos-sidecar" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.702301 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="thanos-sidecar" Nov 24 14:36:56 crc kubenswrapper[4822]: E1124 14:36:56.702315 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="prometheus" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.702321 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="prometheus" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.702482 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="prometheus" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.702500 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="config-reloader" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.702514 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" containerName="thanos-sidecar" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.705409 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.707125 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.707430 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.707632 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-6pdph" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.707802 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.707922 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.708114 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.713654 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.755001 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.792728 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.792785 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.792809 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/51214ce1-190b-4166-9817-af60ba07c73e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.792865 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kmjk\" (UniqueName: \"kubernetes.io/projected/51214ce1-190b-4166-9817-af60ba07c73e-kube-api-access-2kmjk\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.792905 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.792930 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.792966 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/51214ce1-190b-4166-9817-af60ba07c73e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.793001 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.793067 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.793110 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/51214ce1-190b-4166-9817-af60ba07c73e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.793134 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-config\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.876599 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.884898 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894253 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-config\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894300 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894323 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894344 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/51214ce1-190b-4166-9817-af60ba07c73e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894377 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kmjk\" (UniqueName: \"kubernetes.io/projected/51214ce1-190b-4166-9817-af60ba07c73e-kube-api-access-2kmjk\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894412 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894434 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894465 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/51214ce1-190b-4166-9817-af60ba07c73e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894490 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894541 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.894583 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/51214ce1-190b-4166-9817-af60ba07c73e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.898387 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/51214ce1-190b-4166-9817-af60ba07c73e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.898636 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/51214ce1-190b-4166-9817-af60ba07c73e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.901716 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.905378 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-config\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.905861 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.905893 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/af84fe80f7334969b1921a67001ef55fda3e7f45a332f15660d093709721d1df/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.906049 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/51214ce1-190b-4166-9817-af60ba07c73e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.908093 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.904466 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.913139 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.914823 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/51214ce1-190b-4166-9817-af60ba07c73e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.916469 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kmjk\" (UniqueName: \"kubernetes.io/projected/51214ce1-190b-4166-9817-af60ba07c73e-kube-api-access-2kmjk\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.943381 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c165ae10-7797-4222-877f-f5bb6c853296\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c165ae10-7797-4222-877f-f5bb6c853296\") pod \"prometheus-metric-storage-0\" (UID: \"51214ce1-190b-4166-9817-af60ba07c73e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.995333 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdrvf\" (UniqueName: \"kubernetes.io/projected/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-kube-api-access-vdrvf\") pod \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\" (UID: \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.995385 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-operator-scripts\") pod \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\" (UID: \"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.995531 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrllx\" (UniqueName: \"kubernetes.io/projected/53ec312d-2011-41f0-8a13-4268f81bfaa6-kube-api-access-rrllx\") pod \"53ec312d-2011-41f0-8a13-4268f81bfaa6\" (UID: \"53ec312d-2011-41f0-8a13-4268f81bfaa6\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.995637 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec312d-2011-41f0-8a13-4268f81bfaa6-operator-scripts\") pod \"53ec312d-2011-41f0-8a13-4268f81bfaa6\" (UID: \"53ec312d-2011-41f0-8a13-4268f81bfaa6\") " Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.995967 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67" (UID: "ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.996454 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.996494 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53ec312d-2011-41f0-8a13-4268f81bfaa6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53ec312d-2011-41f0-8a13-4268f81bfaa6" (UID: "53ec312d-2011-41f0-8a13-4268f81bfaa6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.998738 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-kube-api-access-vdrvf" (OuterVolumeSpecName: "kube-api-access-vdrvf") pod "ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67" (UID: "ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67"). InnerVolumeSpecName "kube-api-access-vdrvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:56 crc kubenswrapper[4822]: I1124 14:36:56.999035 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53ec312d-2011-41f0-8a13-4268f81bfaa6-kube-api-access-rrllx" (OuterVolumeSpecName: "kube-api-access-rrllx") pod "53ec312d-2011-41f0-8a13-4268f81bfaa6" (UID: "53ec312d-2011-41f0-8a13-4268f81bfaa6"). InnerVolumeSpecName "kube-api-access-rrllx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.050219 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.098534 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrllx\" (UniqueName: \"kubernetes.io/projected/53ec312d-2011-41f0-8a13-4268f81bfaa6-kube-api-access-rrllx\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.099125 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec312d-2011-41f0-8a13-4268f81bfaa6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.099141 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdrvf\" (UniqueName: \"kubernetes.io/projected/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67-kube-api-access-vdrvf\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.332713 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lt4wx" event={"ID":"ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67","Type":"ContainerDied","Data":"1fb051c46c40fe74cf5e577131de00c7c2a18b94fb7484cc84075dc08b91dc06"} Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.332957 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fb051c46c40fe74cf5e577131de00c7c2a18b94fb7484cc84075dc08b91dc06" Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.333005 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lt4wx" Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.341373 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-72ec-account-create-hgstv" event={"ID":"53ec312d-2011-41f0-8a13-4268f81bfaa6","Type":"ContainerDied","Data":"641bbd9e5ecf4de47eae4eedc8a86a18b4d4b7baf032c2e7a50cec0d2a4ab900"} Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.341415 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="641bbd9e5ecf4de47eae4eedc8a86a18b4d4b7baf032c2e7a50cec0d2a4ab900" Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.341493 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-72ec-account-create-hgstv" Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.521872 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:36:57 crc kubenswrapper[4822]: W1124 14:36:57.526666 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51214ce1_190b_4166_9817_af60ba07c73e.slice/crio-b0f84cbe64f94158a07066a94c998037619642de8a1c50b8642b466a2ca9f3c6 WatchSource:0}: Error finding container b0f84cbe64f94158a07066a94c998037619642de8a1c50b8642b466a2ca9f3c6: Status 404 returned error can't find the container with id b0f84cbe64f94158a07066a94c998037619642de8a1c50b8642b466a2ca9f3c6 Nov 24 14:36:57 crc kubenswrapper[4822]: I1124 14:36:57.718078 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08f13b2a-d10d-4b51-be76-b4c510b24d00" path="/var/lib/kubelet/pods/08f13b2a-d10d-4b51-be76-b4c510b24d00/volumes" Nov 24 14:36:58 crc kubenswrapper[4822]: I1124 14:36:58.350041 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"51214ce1-190b-4166-9817-af60ba07c73e","Type":"ContainerStarted","Data":"b0f84cbe64f94158a07066a94c998037619642de8a1c50b8642b466a2ca9f3c6"} Nov 24 14:36:59 crc kubenswrapper[4822]: I1124 14:36:59.546780 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vh28h" podUID="bb7d7a89-bc89-4b18-bbf2-626afb9452b3" containerName="ovn-controller" probeResult="failure" output=< Nov 24 14:36:59 crc kubenswrapper[4822]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 14:36:59 crc kubenswrapper[4822]: > Nov 24 14:37:01 crc kubenswrapper[4822]: I1124 14:37:01.381511 4822 generic.go:334] "Generic (PLEG): container finished" podID="a52d37ed-a00f-4983-88ee-023bd282038e" containerID="276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe" exitCode=0 Nov 24 14:37:01 crc kubenswrapper[4822]: I1124 14:37:01.381591 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a52d37ed-a00f-4983-88ee-023bd282038e","Type":"ContainerDied","Data":"276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe"} Nov 24 14:37:01 crc kubenswrapper[4822]: I1124 14:37:01.383516 4822 generic.go:334] "Generic (PLEG): container finished" podID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" containerID="a11d1a87ad59947fd0c7dafe5c9ace40accaa87c88506ddfd156b36fa2cfb257" exitCode=0 Nov 24 14:37:01 crc kubenswrapper[4822]: I1124 14:37:01.383543 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37","Type":"ContainerDied","Data":"a11d1a87ad59947fd0c7dafe5c9ace40accaa87c88506ddfd156b36fa2cfb257"} Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.574957 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vh28h" podUID="bb7d7a89-bc89-4b18-bbf2-626afb9452b3" containerName="ovn-controller" probeResult="failure" output=< Nov 24 14:37:04 crc kubenswrapper[4822]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 14:37:04 crc kubenswrapper[4822]: > Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.611751 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.617063 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7l6s7" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.878764 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vh28h-config-5c8xb"] Nov 24 14:37:04 crc kubenswrapper[4822]: E1124 14:37:04.879402 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ec312d-2011-41f0-8a13-4268f81bfaa6" containerName="mariadb-account-create" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.879423 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ec312d-2011-41f0-8a13-4268f81bfaa6" containerName="mariadb-account-create" Nov 24 14:37:04 crc kubenswrapper[4822]: E1124 14:37:04.879496 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67" containerName="mariadb-database-create" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.879509 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67" containerName="mariadb-database-create" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.879805 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67" containerName="mariadb-database-create" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.879837 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="53ec312d-2011-41f0-8a13-4268f81bfaa6" containerName="mariadb-account-create" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.880900 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.884054 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.898436 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vh28h-config-5c8xb"] Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.959720 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-scripts\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.960018 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-additional-scripts\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.960149 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgldf\" (UniqueName: \"kubernetes.io/projected/03de2b5d-9a8b-4a20-add8-ce826e3818f2-kube-api-access-hgldf\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.960408 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-log-ovn\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.960553 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run-ovn\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:04 crc kubenswrapper[4822]: I1124 14:37:04.960683 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.063333 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-log-ovn\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.063773 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run-ovn\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.064005 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.064160 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.063770 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-log-ovn\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.063835 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run-ovn\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.064963 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-scripts\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.065161 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-additional-scripts\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.065324 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgldf\" (UniqueName: \"kubernetes.io/projected/03de2b5d-9a8b-4a20-add8-ce826e3818f2-kube-api-access-hgldf\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.066390 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-additional-scripts\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.069882 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-scripts\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.106170 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgldf\" (UniqueName: \"kubernetes.io/projected/03de2b5d-9a8b-4a20-add8-ce826e3818f2-kube-api-access-hgldf\") pod \"ovn-controller-vh28h-config-5c8xb\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:05 crc kubenswrapper[4822]: I1124 14:37:05.207063 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.091530 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vh28h-config-5c8xb"] Nov 24 14:37:06 crc kubenswrapper[4822]: W1124 14:37:06.108655 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03de2b5d_9a8b_4a20_add8_ce826e3818f2.slice/crio-70eaafb124f5729a6a8d4430e8e2a8c600f528dfe52ebe251bed370188ad4f05 WatchSource:0}: Error finding container 70eaafb124f5729a6a8d4430e8e2a8c600f528dfe52ebe251bed370188ad4f05: Status 404 returned error can't find the container with id 70eaafb124f5729a6a8d4430e8e2a8c600f528dfe52ebe251bed370188ad4f05 Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.388711 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="849f7f3c-3605-4e1b-8044-5e063841745e" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.431995 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vh28h-config-5c8xb" event={"ID":"03de2b5d-9a8b-4a20-add8-ce826e3818f2","Type":"ContainerStarted","Data":"f417c33072a5a4a65267ea26ba787c959ddfcb24d817afa0933cd5bf9cd1d067"} Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.432225 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vh28h-config-5c8xb" event={"ID":"03de2b5d-9a8b-4a20-add8-ce826e3818f2","Type":"ContainerStarted","Data":"70eaafb124f5729a6a8d4430e8e2a8c600f528dfe52ebe251bed370188ad4f05"} Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.436540 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a52d37ed-a00f-4983-88ee-023bd282038e","Type":"ContainerStarted","Data":"48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963"} Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.436808 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.439363 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37","Type":"ContainerStarted","Data":"dec06a7f92f37aaabc29aea08f1da7e2195f934ec84abde199e37f03bf023343"} Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.439539 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.442263 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bwmcx" event={"ID":"a39cbd81-d3b4-46a5-a513-031aca8e43c3","Type":"ContainerStarted","Data":"6f803271be52efcd62812b152e9ed55bc4a83f39c3abf960780134527d6744a9"} Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.450884 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vh28h-config-5c8xb" podStartSLOduration=2.450867432 podStartE2EDuration="2.450867432s" podCreationTimestamp="2025-11-24 14:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:37:06.447825183 +0000 UTC m=+1063.564465660" watchObservedRunningTime="2025-11-24 14:37:06.450867432 +0000 UTC m=+1063.567507929" Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.485102 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=55.74274462 podStartE2EDuration="1m7.485085895s" podCreationTimestamp="2025-11-24 14:35:59 +0000 UTC" firstStartedPulling="2025-11-24 14:36:12.043430538 +0000 UTC m=+1009.160071015" lastFinishedPulling="2025-11-24 14:36:23.785771803 +0000 UTC m=+1020.902412290" observedRunningTime="2025-11-24 14:37:06.481817119 +0000 UTC m=+1063.598457606" watchObservedRunningTime="2025-11-24 14:37:06.485085895 +0000 UTC m=+1063.601726372" Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.506984 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-bwmcx" podStartSLOduration=2.098628029 podStartE2EDuration="17.506966396s" podCreationTimestamp="2025-11-24 14:36:49 +0000 UTC" firstStartedPulling="2025-11-24 14:36:50.252150181 +0000 UTC m=+1047.368790658" lastFinishedPulling="2025-11-24 14:37:05.660488548 +0000 UTC m=+1062.777129025" observedRunningTime="2025-11-24 14:37:06.495820559 +0000 UTC m=+1063.612461056" watchObservedRunningTime="2025-11-24 14:37:06.506966396 +0000 UTC m=+1063.623606873" Nov 24 14:37:06 crc kubenswrapper[4822]: I1124 14:37:06.524476 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=58.434504767 podStartE2EDuration="1m7.524451357s" podCreationTimestamp="2025-11-24 14:35:59 +0000 UTC" firstStartedPulling="2025-11-24 14:36:12.721451811 +0000 UTC m=+1009.838092298" lastFinishedPulling="2025-11-24 14:36:21.811398411 +0000 UTC m=+1018.928038888" observedRunningTime="2025-11-24 14:37:06.51803139 +0000 UTC m=+1063.634671887" watchObservedRunningTime="2025-11-24 14:37:06.524451357 +0000 UTC m=+1063.641091834" Nov 24 14:37:07 crc kubenswrapper[4822]: I1124 14:37:07.453124 4822 generic.go:334] "Generic (PLEG): container finished" podID="03de2b5d-9a8b-4a20-add8-ce826e3818f2" containerID="f417c33072a5a4a65267ea26ba787c959ddfcb24d817afa0933cd5bf9cd1d067" exitCode=0 Nov 24 14:37:07 crc kubenswrapper[4822]: I1124 14:37:07.453232 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vh28h-config-5c8xb" event={"ID":"03de2b5d-9a8b-4a20-add8-ce826e3818f2","Type":"ContainerDied","Data":"f417c33072a5a4a65267ea26ba787c959ddfcb24d817afa0933cd5bf9cd1d067"} Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.816091 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.841608 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-log-ovn\") pod \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.841898 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run\") pod \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.841734 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "03de2b5d-9a8b-4a20-add8-ce826e3818f2" (UID: "03de2b5d-9a8b-4a20-add8-ce826e3818f2"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.841953 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-scripts\") pod \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842014 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run" (OuterVolumeSpecName: "var-run") pod "03de2b5d-9a8b-4a20-add8-ce826e3818f2" (UID: "03de2b5d-9a8b-4a20-add8-ce826e3818f2"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842033 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run-ovn\") pod \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842144 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-additional-scripts\") pod \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842215 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgldf\" (UniqueName: \"kubernetes.io/projected/03de2b5d-9a8b-4a20-add8-ce826e3818f2-kube-api-access-hgldf\") pod \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\" (UID: \"03de2b5d-9a8b-4a20-add8-ce826e3818f2\") " Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842221 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "03de2b5d-9a8b-4a20-add8-ce826e3818f2" (UID: "03de2b5d-9a8b-4a20-add8-ce826e3818f2"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842621 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842689 4822 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842698 4822 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842706 4822 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/03de2b5d-9a8b-4a20-add8-ce826e3818f2-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842751 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "03de2b5d-9a8b-4a20-add8-ce826e3818f2" (UID: "03de2b5d-9a8b-4a20-add8-ce826e3818f2"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:08 crc kubenswrapper[4822]: E1124 14:37:08.842798 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:37:08 crc kubenswrapper[4822]: E1124 14:37:08.842811 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:37:08 crc kubenswrapper[4822]: E1124 14:37:08.842850 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:37:40.842836844 +0000 UTC m=+1097.959477321 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.842983 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-scripts" (OuterVolumeSpecName: "scripts") pod "03de2b5d-9a8b-4a20-add8-ce826e3818f2" (UID: "03de2b5d-9a8b-4a20-add8-ce826e3818f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.869930 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03de2b5d-9a8b-4a20-add8-ce826e3818f2-kube-api-access-hgldf" (OuterVolumeSpecName: "kube-api-access-hgldf") pod "03de2b5d-9a8b-4a20-add8-ce826e3818f2" (UID: "03de2b5d-9a8b-4a20-add8-ce826e3818f2"). InnerVolumeSpecName "kube-api-access-hgldf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.947843 4822 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.947972 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgldf\" (UniqueName: \"kubernetes.io/projected/03de2b5d-9a8b-4a20-add8-ce826e3818f2-kube-api-access-hgldf\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:08 crc kubenswrapper[4822]: I1124 14:37:08.948067 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03de2b5d-9a8b-4a20-add8-ce826e3818f2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:09 crc kubenswrapper[4822]: I1124 14:37:09.181597 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vh28h-config-5c8xb"] Nov 24 14:37:09 crc kubenswrapper[4822]: I1124 14:37:09.192029 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vh28h-config-5c8xb"] Nov 24 14:37:09 crc kubenswrapper[4822]: I1124 14:37:09.479700 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vh28h-config-5c8xb" Nov 24 14:37:09 crc kubenswrapper[4822]: I1124 14:37:09.479713 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70eaafb124f5729a6a8d4430e8e2a8c600f528dfe52ebe251bed370188ad4f05" Nov 24 14:37:09 crc kubenswrapper[4822]: I1124 14:37:09.482625 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"51214ce1-190b-4166-9817-af60ba07c73e","Type":"ContainerStarted","Data":"c82069e8988c95a1a03499284f4c4968ebe09aa18983aa78cc27dbb8dc920fcf"} Nov 24 14:37:09 crc kubenswrapper[4822]: I1124 14:37:09.550188 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-vh28h" Nov 24 14:37:09 crc kubenswrapper[4822]: I1124 14:37:09.721009 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03de2b5d-9a8b-4a20-add8-ce826e3818f2" path="/var/lib/kubelet/pods/03de2b5d-9a8b-4a20-add8-ce826e3818f2/volumes" Nov 24 14:37:12 crc kubenswrapper[4822]: I1124 14:37:12.512466 4822 generic.go:334] "Generic (PLEG): container finished" podID="a39cbd81-d3b4-46a5-a513-031aca8e43c3" containerID="6f803271be52efcd62812b152e9ed55bc4a83f39c3abf960780134527d6744a9" exitCode=0 Nov 24 14:37:12 crc kubenswrapper[4822]: I1124 14:37:12.512575 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bwmcx" event={"ID":"a39cbd81-d3b4-46a5-a513-031aca8e43c3","Type":"ContainerDied","Data":"6f803271be52efcd62812b152e9ed55bc4a83f39c3abf960780134527d6744a9"} Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.354171 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bwmcx" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.365377 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-db-sync-config-data\") pod \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.365447 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjsvq\" (UniqueName: \"kubernetes.io/projected/a39cbd81-d3b4-46a5-a513-031aca8e43c3-kube-api-access-sjsvq\") pod \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.372024 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a39cbd81-d3b4-46a5-a513-031aca8e43c3" (UID: "a39cbd81-d3b4-46a5-a513-031aca8e43c3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.374336 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a39cbd81-d3b4-46a5-a513-031aca8e43c3-kube-api-access-sjsvq" (OuterVolumeSpecName: "kube-api-access-sjsvq") pod "a39cbd81-d3b4-46a5-a513-031aca8e43c3" (UID: "a39cbd81-d3b4-46a5-a513-031aca8e43c3"). InnerVolumeSpecName "kube-api-access-sjsvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.467861 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-combined-ca-bundle\") pod \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.468172 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-config-data\") pod \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\" (UID: \"a39cbd81-d3b4-46a5-a513-031aca8e43c3\") " Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.468870 4822 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.468915 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjsvq\" (UniqueName: \"kubernetes.io/projected/a39cbd81-d3b4-46a5-a513-031aca8e43c3-kube-api-access-sjsvq\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.489463 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a39cbd81-d3b4-46a5-a513-031aca8e43c3" (UID: "a39cbd81-d3b4-46a5-a513-031aca8e43c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.537271 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-config-data" (OuterVolumeSpecName: "config-data") pod "a39cbd81-d3b4-46a5-a513-031aca8e43c3" (UID: "a39cbd81-d3b4-46a5-a513-031aca8e43c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.546010 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bwmcx" event={"ID":"a39cbd81-d3b4-46a5-a513-031aca8e43c3","Type":"ContainerDied","Data":"51db94409f902bc660472c31e47f64d6d9f70b908f447bab4b8276c96537b1c3"} Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.546080 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51db94409f902bc660472c31e47f64d6d9f70b908f447bab4b8276c96537b1c3" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.546106 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bwmcx" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.571681 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.571902 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a39cbd81-d3b4-46a5-a513-031aca8e43c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.971591 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-gkssz"] Nov 24 14:37:14 crc kubenswrapper[4822]: E1124 14:37:14.972385 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03de2b5d-9a8b-4a20-add8-ce826e3818f2" containerName="ovn-config" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.972412 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="03de2b5d-9a8b-4a20-add8-ce826e3818f2" containerName="ovn-config" Nov 24 14:37:14 crc kubenswrapper[4822]: E1124 14:37:14.972442 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a39cbd81-d3b4-46a5-a513-031aca8e43c3" containerName="glance-db-sync" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.972452 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a39cbd81-d3b4-46a5-a513-031aca8e43c3" containerName="glance-db-sync" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.972700 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a39cbd81-d3b4-46a5-a513-031aca8e43c3" containerName="glance-db-sync" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.972722 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="03de2b5d-9a8b-4a20-add8-ce826e3818f2" containerName="ovn-config" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.973955 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:14 crc kubenswrapper[4822]: I1124 14:37:14.980993 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-gkssz"] Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.081776 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-config\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.081841 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.081918 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flxjh\" (UniqueName: \"kubernetes.io/projected/981326ac-656e-410e-8016-7869af8ed9ea-kube-api-access-flxjh\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.081970 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.082111 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.184022 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-config\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.184302 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.184389 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flxjh\" (UniqueName: \"kubernetes.io/projected/981326ac-656e-410e-8016-7869af8ed9ea-kube-api-access-flxjh\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.184481 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.184620 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.185117 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.185232 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-config\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.185540 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.185994 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.200142 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flxjh\" (UniqueName: \"kubernetes.io/projected/981326ac-656e-410e-8016-7869af8ed9ea-kube-api-access-flxjh\") pod \"dnsmasq-dns-5b946c75cc-gkssz\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.289895 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:15 crc kubenswrapper[4822]: I1124 14:37:15.779417 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-gkssz"] Nov 24 14:37:16 crc kubenswrapper[4822]: I1124 14:37:16.388034 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="849f7f3c-3605-4e1b-8044-5e063841745e" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 14:37:16 crc kubenswrapper[4822]: I1124 14:37:16.573940 4822 generic.go:334] "Generic (PLEG): container finished" podID="981326ac-656e-410e-8016-7869af8ed9ea" containerID="99bc53f4c11169da9d920d934dbbce72d43150e8cc269751b59db4fa9c0e621d" exitCode=0 Nov 24 14:37:16 crc kubenswrapper[4822]: I1124 14:37:16.574011 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" event={"ID":"981326ac-656e-410e-8016-7869af8ed9ea","Type":"ContainerDied","Data":"99bc53f4c11169da9d920d934dbbce72d43150e8cc269751b59db4fa9c0e621d"} Nov 24 14:37:16 crc kubenswrapper[4822]: I1124 14:37:16.574086 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" event={"ID":"981326ac-656e-410e-8016-7869af8ed9ea","Type":"ContainerStarted","Data":"f49ebb866eb886373277ffc7081dae0b33d8680b3f2378554f89ed3b471d7a7c"} Nov 24 14:37:17 crc kubenswrapper[4822]: I1124 14:37:17.587498 4822 generic.go:334] "Generic (PLEG): container finished" podID="51214ce1-190b-4166-9817-af60ba07c73e" containerID="c82069e8988c95a1a03499284f4c4968ebe09aa18983aa78cc27dbb8dc920fcf" exitCode=0 Nov 24 14:37:17 crc kubenswrapper[4822]: I1124 14:37:17.588017 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"51214ce1-190b-4166-9817-af60ba07c73e","Type":"ContainerDied","Data":"c82069e8988c95a1a03499284f4c4968ebe09aa18983aa78cc27dbb8dc920fcf"} Nov 24 14:37:17 crc kubenswrapper[4822]: I1124 14:37:17.600027 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" event={"ID":"981326ac-656e-410e-8016-7869af8ed9ea","Type":"ContainerStarted","Data":"b77a3bdca919e625a1558a04b75e41b3591c9a2804450f405b7ab304139e6a2c"} Nov 24 14:37:17 crc kubenswrapper[4822]: I1124 14:37:17.600881 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:17 crc kubenswrapper[4822]: I1124 14:37:17.719455 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" podStartSLOduration=3.719437534 podStartE2EDuration="3.719437534s" podCreationTimestamp="2025-11-24 14:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:37:17.716584471 +0000 UTC m=+1074.833224948" watchObservedRunningTime="2025-11-24 14:37:17.719437534 +0000 UTC m=+1074.836078011" Nov 24 14:37:18 crc kubenswrapper[4822]: I1124 14:37:18.612830 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"51214ce1-190b-4166-9817-af60ba07c73e","Type":"ContainerStarted","Data":"5178e2e3d9f5920d33be254e355a5f8f405cd2c27209e140aab9ff488e225a8a"} Nov 24 14:37:20 crc kubenswrapper[4822]: I1124 14:37:20.727482 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 14:37:20 crc kubenswrapper[4822]: I1124 14:37:20.783111 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.185462 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-lhp86"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.187127 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.194863 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lhp86"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.301883 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5f21-account-create-ds2h6"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.303239 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.308731 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-bhnlm"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.308967 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.310033 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.315066 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae75d4f-b139-4b12-b581-3362d46de89e-operator-scripts\") pod \"cinder-db-create-lhp86\" (UID: \"cae75d4f-b139-4b12-b581-3362d46de89e\") " pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.315258 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfrnx\" (UniqueName: \"kubernetes.io/projected/cae75d4f-b139-4b12-b581-3362d46de89e-kube-api-access-vfrnx\") pod \"cinder-db-create-lhp86\" (UID: \"cae75d4f-b139-4b12-b581-3362d46de89e\") " pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.317556 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5f21-account-create-ds2h6"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.327273 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bhnlm"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.411310 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8fc4-account-create-tgmfq"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.412612 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.414363 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.416381 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfrnx\" (UniqueName: \"kubernetes.io/projected/cae75d4f-b139-4b12-b581-3362d46de89e-kube-api-access-vfrnx\") pod \"cinder-db-create-lhp86\" (UID: \"cae75d4f-b139-4b12-b581-3362d46de89e\") " pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.416413 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pnlx\" (UniqueName: \"kubernetes.io/projected/728c9997-66cf-4996-8b4d-75af0e3ba55c-kube-api-access-7pnlx\") pod \"cinder-5f21-account-create-ds2h6\" (UID: \"728c9997-66cf-4996-8b4d-75af0e3ba55c\") " pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.416459 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/908b60c2-798e-4feb-90da-4cd08d479578-operator-scripts\") pod \"barbican-db-create-bhnlm\" (UID: \"908b60c2-798e-4feb-90da-4cd08d479578\") " pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.416493 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glppg\" (UniqueName: \"kubernetes.io/projected/908b60c2-798e-4feb-90da-4cd08d479578-kube-api-access-glppg\") pod \"barbican-db-create-bhnlm\" (UID: \"908b60c2-798e-4feb-90da-4cd08d479578\") " pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.416515 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae75d4f-b139-4b12-b581-3362d46de89e-operator-scripts\") pod \"cinder-db-create-lhp86\" (UID: \"cae75d4f-b139-4b12-b581-3362d46de89e\") " pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.416538 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/728c9997-66cf-4996-8b4d-75af0e3ba55c-operator-scripts\") pod \"cinder-5f21-account-create-ds2h6\" (UID: \"728c9997-66cf-4996-8b4d-75af0e3ba55c\") " pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.417456 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae75d4f-b139-4b12-b581-3362d46de89e-operator-scripts\") pod \"cinder-db-create-lhp86\" (UID: \"cae75d4f-b139-4b12-b581-3362d46de89e\") " pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.417609 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8fc4-account-create-tgmfq"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.437914 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfrnx\" (UniqueName: \"kubernetes.io/projected/cae75d4f-b139-4b12-b581-3362d46de89e-kube-api-access-vfrnx\") pod \"cinder-db-create-lhp86\" (UID: \"cae75d4f-b139-4b12-b581-3362d46de89e\") " pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.487567 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-57687"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.488699 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.491513 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.492385 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r2n9r" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.492507 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.493524 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.503242 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-57687"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.513565 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.517879 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/908b60c2-798e-4feb-90da-4cd08d479578-operator-scripts\") pod \"barbican-db-create-bhnlm\" (UID: \"908b60c2-798e-4feb-90da-4cd08d479578\") " pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.517946 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnxjz\" (UniqueName: \"kubernetes.io/projected/2fc8a762-b8f7-4246-aaec-d6670cf715a5-kube-api-access-rnxjz\") pod \"barbican-8fc4-account-create-tgmfq\" (UID: \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\") " pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.517976 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glppg\" (UniqueName: \"kubernetes.io/projected/908b60c2-798e-4feb-90da-4cd08d479578-kube-api-access-glppg\") pod \"barbican-db-create-bhnlm\" (UID: \"908b60c2-798e-4feb-90da-4cd08d479578\") " pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.518021 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/728c9997-66cf-4996-8b4d-75af0e3ba55c-operator-scripts\") pod \"cinder-5f21-account-create-ds2h6\" (UID: \"728c9997-66cf-4996-8b4d-75af0e3ba55c\") " pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.518199 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fc8a762-b8f7-4246-aaec-d6670cf715a5-operator-scripts\") pod \"barbican-8fc4-account-create-tgmfq\" (UID: \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\") " pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.518348 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pnlx\" (UniqueName: \"kubernetes.io/projected/728c9997-66cf-4996-8b4d-75af0e3ba55c-kube-api-access-7pnlx\") pod \"cinder-5f21-account-create-ds2h6\" (UID: \"728c9997-66cf-4996-8b4d-75af0e3ba55c\") " pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.518794 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/908b60c2-798e-4feb-90da-4cd08d479578-operator-scripts\") pod \"barbican-db-create-bhnlm\" (UID: \"908b60c2-798e-4feb-90da-4cd08d479578\") " pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.519358 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/728c9997-66cf-4996-8b4d-75af0e3ba55c-operator-scripts\") pod \"cinder-5f21-account-create-ds2h6\" (UID: \"728c9997-66cf-4996-8b4d-75af0e3ba55c\") " pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.542903 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glppg\" (UniqueName: \"kubernetes.io/projected/908b60c2-798e-4feb-90da-4cd08d479578-kube-api-access-glppg\") pod \"barbican-db-create-bhnlm\" (UID: \"908b60c2-798e-4feb-90da-4cd08d479578\") " pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.547992 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pnlx\" (UniqueName: \"kubernetes.io/projected/728c9997-66cf-4996-8b4d-75af0e3ba55c-kube-api-access-7pnlx\") pod \"cinder-5f21-account-create-ds2h6\" (UID: \"728c9997-66cf-4996-8b4d-75af0e3ba55c\") " pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.596643 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-create-ltdt2"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.597797 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.611532 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-ltdt2"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.619725 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-config-data\") pod \"keystone-db-sync-57687\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.619793 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fc8a762-b8f7-4246-aaec-d6670cf715a5-operator-scripts\") pod \"barbican-8fc4-account-create-tgmfq\" (UID: \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\") " pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.619831 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-combined-ca-bundle\") pod \"keystone-db-sync-57687\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.619872 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnxjz\" (UniqueName: \"kubernetes.io/projected/2fc8a762-b8f7-4246-aaec-d6670cf715a5-kube-api-access-rnxjz\") pod \"barbican-8fc4-account-create-tgmfq\" (UID: \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\") " pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.619915 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc7cn\" (UniqueName: \"kubernetes.io/projected/274d41ac-17fe-48ef-ad92-f769c96b219b-kube-api-access-rc7cn\") pod \"keystone-db-sync-57687\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.620713 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fc8a762-b8f7-4246-aaec-d6670cf715a5-operator-scripts\") pod \"barbican-8fc4-account-create-tgmfq\" (UID: \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\") " pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.622295 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.635664 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.636739 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnxjz\" (UniqueName: \"kubernetes.io/projected/2fc8a762-b8f7-4246-aaec-d6670cf715a5-kube-api-access-rnxjz\") pod \"barbican-8fc4-account-create-tgmfq\" (UID: \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\") " pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.721386 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc7cn\" (UniqueName: \"kubernetes.io/projected/274d41ac-17fe-48ef-ad92-f769c96b219b-kube-api-access-rc7cn\") pod \"keystone-db-sync-57687\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.721738 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8rsp\" (UniqueName: \"kubernetes.io/projected/5254b05c-78e8-49fa-932b-4f36b5fb907d-kube-api-access-s8rsp\") pod \"cloudkitty-db-create-ltdt2\" (UID: \"5254b05c-78e8-49fa-932b-4f36b5fb907d\") " pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.721763 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-config-data\") pod \"keystone-db-sync-57687\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.721779 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5254b05c-78e8-49fa-932b-4f36b5fb907d-operator-scripts\") pod \"cloudkitty-db-create-ltdt2\" (UID: \"5254b05c-78e8-49fa-932b-4f36b5fb907d\") " pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.721857 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-combined-ca-bundle\") pod \"keystone-db-sync-57687\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.735830 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.736513 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-config-data\") pod \"keystone-db-sync-57687\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.736815 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-combined-ca-bundle\") pod \"keystone-db-sync-57687\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.751553 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-3aed-account-create-db4jq"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.752638 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.755008 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-db-secret" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.768723 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-4kdxz"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.769897 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.798398 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc7cn\" (UniqueName: \"kubernetes.io/projected/274d41ac-17fe-48ef-ad92-f769c96b219b-kube-api-access-rc7cn\") pod \"keystone-db-sync-57687\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.809549 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-57687" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.824012 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmt9r\" (UniqueName: \"kubernetes.io/projected/a1c97044-4916-4bdd-aae7-33069d43663e-kube-api-access-rmt9r\") pod \"cloudkitty-3aed-account-create-db4jq\" (UID: \"a1c97044-4916-4bdd-aae7-33069d43663e\") " pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.824146 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c97044-4916-4bdd-aae7-33069d43663e-operator-scripts\") pod \"cloudkitty-3aed-account-create-db4jq\" (UID: \"a1c97044-4916-4bdd-aae7-33069d43663e\") " pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.849261 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-3aed-account-create-db4jq"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.855505 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8rsp\" (UniqueName: \"kubernetes.io/projected/5254b05c-78e8-49fa-932b-4f36b5fb907d-kube-api-access-s8rsp\") pod \"cloudkitty-db-create-ltdt2\" (UID: \"5254b05c-78e8-49fa-932b-4f36b5fb907d\") " pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.855577 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5254b05c-78e8-49fa-932b-4f36b5fb907d-operator-scripts\") pod \"cloudkitty-db-create-ltdt2\" (UID: \"5254b05c-78e8-49fa-932b-4f36b5fb907d\") " pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.857347 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5254b05c-78e8-49fa-932b-4f36b5fb907d-operator-scripts\") pod \"cloudkitty-db-create-ltdt2\" (UID: \"5254b05c-78e8-49fa-932b-4f36b5fb907d\") " pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.884284 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4kdxz"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.905693 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8rsp\" (UniqueName: \"kubernetes.io/projected/5254b05c-78e8-49fa-932b-4f36b5fb907d-kube-api-access-s8rsp\") pod \"cloudkitty-db-create-ltdt2\" (UID: \"5254b05c-78e8-49fa-932b-4f36b5fb907d\") " pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.960262 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-83e5-account-create-ttzgw"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.961889 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.964003 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdk89\" (UniqueName: \"kubernetes.io/projected/bc04d3ff-c5b1-4134-b544-ca9341af72b4-kube-api-access-hdk89\") pod \"neutron-db-create-4kdxz\" (UID: \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\") " pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.964067 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmt9r\" (UniqueName: \"kubernetes.io/projected/a1c97044-4916-4bdd-aae7-33069d43663e-kube-api-access-rmt9r\") pod \"cloudkitty-3aed-account-create-db4jq\" (UID: \"a1c97044-4916-4bdd-aae7-33069d43663e\") " pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.964159 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc04d3ff-c5b1-4134-b544-ca9341af72b4-operator-scripts\") pod \"neutron-db-create-4kdxz\" (UID: \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\") " pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.964220 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c97044-4916-4bdd-aae7-33069d43663e-operator-scripts\") pod \"cloudkitty-3aed-account-create-db4jq\" (UID: \"a1c97044-4916-4bdd-aae7-33069d43663e\") " pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.965048 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c97044-4916-4bdd-aae7-33069d43663e-operator-scripts\") pod \"cloudkitty-3aed-account-create-db4jq\" (UID: \"a1c97044-4916-4bdd-aae7-33069d43663e\") " pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.965448 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.973272 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-83e5-account-create-ttzgw"] Nov 24 14:37:21 crc kubenswrapper[4822]: I1124 14:37:21.984881 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmt9r\" (UniqueName: \"kubernetes.io/projected/a1c97044-4916-4bdd-aae7-33069d43663e-kube-api-access-rmt9r\") pod \"cloudkitty-3aed-account-create-db4jq\" (UID: \"a1c97044-4916-4bdd-aae7-33069d43663e\") " pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.065701 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ddqt\" (UniqueName: \"kubernetes.io/projected/32b5e812-c653-4437-ae27-b7a1635085a6-kube-api-access-6ddqt\") pod \"neutron-83e5-account-create-ttzgw\" (UID: \"32b5e812-c653-4437-ae27-b7a1635085a6\") " pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.065756 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdk89\" (UniqueName: \"kubernetes.io/projected/bc04d3ff-c5b1-4134-b544-ca9341af72b4-kube-api-access-hdk89\") pod \"neutron-db-create-4kdxz\" (UID: \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\") " pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.065816 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b5e812-c653-4437-ae27-b7a1635085a6-operator-scripts\") pod \"neutron-83e5-account-create-ttzgw\" (UID: \"32b5e812-c653-4437-ae27-b7a1635085a6\") " pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.065858 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc04d3ff-c5b1-4134-b544-ca9341af72b4-operator-scripts\") pod \"neutron-db-create-4kdxz\" (UID: \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\") " pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.066596 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc04d3ff-c5b1-4134-b544-ca9341af72b4-operator-scripts\") pod \"neutron-db-create-4kdxz\" (UID: \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\") " pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.083868 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdk89\" (UniqueName: \"kubernetes.io/projected/bc04d3ff-c5b1-4134-b544-ca9341af72b4-kube-api-access-hdk89\") pod \"neutron-db-create-4kdxz\" (UID: \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\") " pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.133965 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.145863 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.167804 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b5e812-c653-4437-ae27-b7a1635085a6-operator-scripts\") pod \"neutron-83e5-account-create-ttzgw\" (UID: \"32b5e812-c653-4437-ae27-b7a1635085a6\") " pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.167959 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ddqt\" (UniqueName: \"kubernetes.io/projected/32b5e812-c653-4437-ae27-b7a1635085a6-kube-api-access-6ddqt\") pod \"neutron-83e5-account-create-ttzgw\" (UID: \"32b5e812-c653-4437-ae27-b7a1635085a6\") " pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.168378 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.168732 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b5e812-c653-4437-ae27-b7a1635085a6-operator-scripts\") pod \"neutron-83e5-account-create-ttzgw\" (UID: \"32b5e812-c653-4437-ae27-b7a1635085a6\") " pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.190120 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ddqt\" (UniqueName: \"kubernetes.io/projected/32b5e812-c653-4437-ae27-b7a1635085a6-kube-api-access-6ddqt\") pod \"neutron-83e5-account-create-ttzgw\" (UID: \"32b5e812-c653-4437-ae27-b7a1635085a6\") " pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.290458 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.297095 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lhp86"] Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.304507 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bhnlm"] Nov 24 14:37:22 crc kubenswrapper[4822]: W1124 14:37:22.328711 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod908b60c2_798e_4feb_90da_4cd08d479578.slice/crio-38eeb7779e82002497451f647f3c2d893f2cb8fe4e809c842e808c4321adb6f0 WatchSource:0}: Error finding container 38eeb7779e82002497451f647f3c2d893f2cb8fe4e809c842e808c4321adb6f0: Status 404 returned error can't find the container with id 38eeb7779e82002497451f647f3c2d893f2cb8fe4e809c842e808c4321adb6f0 Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.406858 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5f21-account-create-ds2h6"] Nov 24 14:37:22 crc kubenswrapper[4822]: W1124 14:37:22.454517 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod728c9997_66cf_4996_8b4d_75af0e3ba55c.slice/crio-fa30fe893432cdf707216cba5f48ceb8041682db66d3494d4c1a0526e81e5585 WatchSource:0}: Error finding container fa30fe893432cdf707216cba5f48ceb8041682db66d3494d4c1a0526e81e5585: Status 404 returned error can't find the container with id fa30fe893432cdf707216cba5f48ceb8041682db66d3494d4c1a0526e81e5585 Nov 24 14:37:22 crc kubenswrapper[4822]: I1124 14:37:22.534635 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-57687"] Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.577743 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8fc4-account-create-tgmfq"] Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.660977 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bhnlm" event={"ID":"908b60c2-798e-4feb-90da-4cd08d479578","Type":"ContainerStarted","Data":"38eeb7779e82002497451f647f3c2d893f2cb8fe4e809c842e808c4321adb6f0"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.661791 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5f21-account-create-ds2h6" event={"ID":"728c9997-66cf-4996-8b4d-75af0e3ba55c","Type":"ContainerStarted","Data":"fa30fe893432cdf707216cba5f48ceb8041682db66d3494d4c1a0526e81e5585"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.662473 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-57687" event={"ID":"274d41ac-17fe-48ef-ad92-f769c96b219b","Type":"ContainerStarted","Data":"7a05ae2f1e523c8a96f5f349a2166f0d6c795dbf1dff11f9c88137d4b27fcbb2"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.663134 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lhp86" event={"ID":"cae75d4f-b139-4b12-b581-3362d46de89e","Type":"ContainerStarted","Data":"a4778d288dca17ad1d3de324e73393ab2905d4dd4ffcedf43ff1c5e3ef82a4bb"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.669115 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8fc4-account-create-tgmfq" event={"ID":"2fc8a762-b8f7-4246-aaec-d6670cf715a5","Type":"ContainerStarted","Data":"24b41a29acc9a11d2e68a4be013da9cb89289950e3e9f34d6a070641630772c7"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.677541 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"51214ce1-190b-4166-9817-af60ba07c73e","Type":"ContainerStarted","Data":"55902e180753fea7c51ca346fe404bf33a0197fa5c30bb394d5ebd15c3bd84dd"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.677580 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"51214ce1-190b-4166-9817-af60ba07c73e","Type":"ContainerStarted","Data":"b41870dcc6480b499dc52cee4bd518f3b1f699bcce784f81a15ccaa8ca310dde"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.707621 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=26.707602382 podStartE2EDuration="26.707602382s" podCreationTimestamp="2025-11-24 14:36:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:37:22.702777351 +0000 UTC m=+1079.819417838" watchObservedRunningTime="2025-11-24 14:37:22.707602382 +0000 UTC m=+1079.824242859" Nov 24 14:37:23 crc kubenswrapper[4822]: W1124 14:37:22.746611 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5254b05c_78e8_49fa_932b_4f36b5fb907d.slice/crio-1ad573ed4dd23c5f53ed7fcb297baf46ad76fc0b257cd38e8feba6e50d037ad4 WatchSource:0}: Error finding container 1ad573ed4dd23c5f53ed7fcb297baf46ad76fc0b257cd38e8feba6e50d037ad4: Status 404 returned error can't find the container with id 1ad573ed4dd23c5f53ed7fcb297baf46ad76fc0b257cd38e8feba6e50d037ad4 Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.747273 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-ltdt2"] Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:22.868224 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-3aed-account-create-db4jq"] Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.525709 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4kdxz"] Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.551258 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-83e5-account-create-ttzgw"] Nov 24 14:37:23 crc kubenswrapper[4822]: W1124 14:37:23.578308 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc04d3ff_c5b1_4134_b544_ca9341af72b4.slice/crio-df2bae0321c9376a4d5408425737723f4478555abad84bf2c3aea731e143f2a1 WatchSource:0}: Error finding container df2bae0321c9376a4d5408425737723f4478555abad84bf2c3aea731e143f2a1: Status 404 returned error can't find the container with id df2bae0321c9376a4d5408425737723f4478555abad84bf2c3aea731e143f2a1 Nov 24 14:37:23 crc kubenswrapper[4822]: W1124 14:37:23.578607 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32b5e812_c653_4437_ae27_b7a1635085a6.slice/crio-b7fa3e4a4b44bf175fa8057abc34dfa722a82914ce546d01bd12494dd728fe94 WatchSource:0}: Error finding container b7fa3e4a4b44bf175fa8057abc34dfa722a82914ce546d01bd12494dd728fe94: Status 404 returned error can't find the container with id b7fa3e4a4b44bf175fa8057abc34dfa722a82914ce546d01bd12494dd728fe94 Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.688591 4822 generic.go:334] "Generic (PLEG): container finished" podID="908b60c2-798e-4feb-90da-4cd08d479578" containerID="3fa43b7ab86ca72cd1e37740e9b76218c196ac8152192ac8a570513ea72f9242" exitCode=0 Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.688646 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bhnlm" event={"ID":"908b60c2-798e-4feb-90da-4cd08d479578","Type":"ContainerDied","Data":"3fa43b7ab86ca72cd1e37740e9b76218c196ac8152192ac8a570513ea72f9242"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.690678 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-83e5-account-create-ttzgw" event={"ID":"32b5e812-c653-4437-ae27-b7a1635085a6","Type":"ContainerStarted","Data":"b7fa3e4a4b44bf175fa8057abc34dfa722a82914ce546d01bd12494dd728fe94"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.695365 4822 generic.go:334] "Generic (PLEG): container finished" podID="728c9997-66cf-4996-8b4d-75af0e3ba55c" containerID="e94a0fac2144c591a1d37f88873d350885e2f13e7467396e159a9984f10a8c05" exitCode=0 Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.695421 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5f21-account-create-ds2h6" event={"ID":"728c9997-66cf-4996-8b4d-75af0e3ba55c","Type":"ContainerDied","Data":"e94a0fac2144c591a1d37f88873d350885e2f13e7467396e159a9984f10a8c05"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.696589 4822 generic.go:334] "Generic (PLEG): container finished" podID="5254b05c-78e8-49fa-932b-4f36b5fb907d" containerID="3d2c9ab3c17bdf612aae63973f64a30d64aefbdb503d7e6ad009200153c7bf7e" exitCode=0 Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.696626 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-ltdt2" event={"ID":"5254b05c-78e8-49fa-932b-4f36b5fb907d","Type":"ContainerDied","Data":"3d2c9ab3c17bdf612aae63973f64a30d64aefbdb503d7e6ad009200153c7bf7e"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.696642 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-ltdt2" event={"ID":"5254b05c-78e8-49fa-932b-4f36b5fb907d","Type":"ContainerStarted","Data":"1ad573ed4dd23c5f53ed7fcb297baf46ad76fc0b257cd38e8feba6e50d037ad4"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.697885 4822 generic.go:334] "Generic (PLEG): container finished" podID="a1c97044-4916-4bdd-aae7-33069d43663e" containerID="ead076bf4ebebd558975431b3a838a550c7480d92a1fa7b2a68615ec0bc34101" exitCode=0 Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.697944 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-3aed-account-create-db4jq" event={"ID":"a1c97044-4916-4bdd-aae7-33069d43663e","Type":"ContainerDied","Data":"ead076bf4ebebd558975431b3a838a550c7480d92a1fa7b2a68615ec0bc34101"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.697967 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-3aed-account-create-db4jq" event={"ID":"a1c97044-4916-4bdd-aae7-33069d43663e","Type":"ContainerStarted","Data":"c2097bd114bfbddbaf73a4eb52f9e30a14f309655d1863a081957de6a1dc26eb"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.698754 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4kdxz" event={"ID":"bc04d3ff-c5b1-4134-b544-ca9341af72b4","Type":"ContainerStarted","Data":"df2bae0321c9376a4d5408425737723f4478555abad84bf2c3aea731e143f2a1"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.699651 4822 generic.go:334] "Generic (PLEG): container finished" podID="cae75d4f-b139-4b12-b581-3362d46de89e" containerID="9fee266f7490312ca4eb42305221c99f63fc4fc82744130d03bc660f88f60949" exitCode=0 Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.699687 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lhp86" event={"ID":"cae75d4f-b139-4b12-b581-3362d46de89e","Type":"ContainerDied","Data":"9fee266f7490312ca4eb42305221c99f63fc4fc82744130d03bc660f88f60949"} Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.701283 4822 generic.go:334] "Generic (PLEG): container finished" podID="2fc8a762-b8f7-4246-aaec-d6670cf715a5" containerID="e7e61c4499846d98aed167fd29eaebb8746aded78421eb4f90753c6cc30bf6fd" exitCode=0 Nov 24 14:37:23 crc kubenswrapper[4822]: I1124 14:37:23.702156 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8fc4-account-create-tgmfq" event={"ID":"2fc8a762-b8f7-4246-aaec-d6670cf715a5","Type":"ContainerDied","Data":"e7e61c4499846d98aed167fd29eaebb8746aded78421eb4f90753c6cc30bf6fd"} Nov 24 14:37:24 crc kubenswrapper[4822]: I1124 14:37:24.716100 4822 generic.go:334] "Generic (PLEG): container finished" podID="bc04d3ff-c5b1-4134-b544-ca9341af72b4" containerID="60b15c160bd848838dc3599ba9bad2bc7838cc8b2ebfd95c2721f7dca97fb6ff" exitCode=0 Nov 24 14:37:24 crc kubenswrapper[4822]: I1124 14:37:24.716172 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4kdxz" event={"ID":"bc04d3ff-c5b1-4134-b544-ca9341af72b4","Type":"ContainerDied","Data":"60b15c160bd848838dc3599ba9bad2bc7838cc8b2ebfd95c2721f7dca97fb6ff"} Nov 24 14:37:24 crc kubenswrapper[4822]: I1124 14:37:24.719255 4822 generic.go:334] "Generic (PLEG): container finished" podID="32b5e812-c653-4437-ae27-b7a1635085a6" containerID="7e65a9dd55dc67b0e7d2c121ce972f4b2dfc73b91d356be0d73b997a699db30c" exitCode=0 Nov 24 14:37:24 crc kubenswrapper[4822]: I1124 14:37:24.719439 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-83e5-account-create-ttzgw" event={"ID":"32b5e812-c653-4437-ae27-b7a1635085a6","Type":"ContainerDied","Data":"7e65a9dd55dc67b0e7d2c121ce972f4b2dfc73b91d356be0d73b997a699db30c"} Nov 24 14:37:25 crc kubenswrapper[4822]: I1124 14:37:25.295723 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:37:25 crc kubenswrapper[4822]: I1124 14:37:25.365533 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-s9xbh"] Nov 24 14:37:25 crc kubenswrapper[4822]: I1124 14:37:25.365770 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-s9xbh" podUID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" containerName="dnsmasq-dns" containerID="cri-o://5df2a9dad76627891e3173a6132be88f4f251a88fcb642e77f4cfb5d40008de3" gracePeriod=10 Nov 24 14:37:25 crc kubenswrapper[4822]: I1124 14:37:25.733377 4822 generic.go:334] "Generic (PLEG): container finished" podID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" containerID="5df2a9dad76627891e3173a6132be88f4f251a88fcb642e77f4cfb5d40008de3" exitCode=0 Nov 24 14:37:25 crc kubenswrapper[4822]: I1124 14:37:25.733450 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-s9xbh" event={"ID":"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176","Type":"ContainerDied","Data":"5df2a9dad76627891e3173a6132be88f4f251a88fcb642e77f4cfb5d40008de3"} Nov 24 14:37:26 crc kubenswrapper[4822]: I1124 14:37:26.168242 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-s9xbh" podUID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Nov 24 14:37:26 crc kubenswrapper[4822]: I1124 14:37:26.384464 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 14:37:27 crc kubenswrapper[4822]: I1124 14:37:27.051374 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 24 14:37:27 crc kubenswrapper[4822]: I1124 14:37:27.051463 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 24 14:37:27 crc kubenswrapper[4822]: I1124 14:37:27.059243 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 24 14:37:27 crc kubenswrapper[4822]: I1124 14:37:27.756097 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.370586 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.402655 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.409591 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.423933 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.440035 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.499344 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.502266 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.508952 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.514805 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnxjz\" (UniqueName: \"kubernetes.io/projected/2fc8a762-b8f7-4246-aaec-d6670cf715a5-kube-api-access-rnxjz\") pod \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\" (UID: \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.514915 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmt9r\" (UniqueName: \"kubernetes.io/projected/a1c97044-4916-4bdd-aae7-33069d43663e-kube-api-access-rmt9r\") pod \"a1c97044-4916-4bdd-aae7-33069d43663e\" (UID: \"a1c97044-4916-4bdd-aae7-33069d43663e\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.515063 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae75d4f-b139-4b12-b581-3362d46de89e-operator-scripts\") pod \"cae75d4f-b139-4b12-b581-3362d46de89e\" (UID: \"cae75d4f-b139-4b12-b581-3362d46de89e\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.515192 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfrnx\" (UniqueName: \"kubernetes.io/projected/cae75d4f-b139-4b12-b581-3362d46de89e-kube-api-access-vfrnx\") pod \"cae75d4f-b139-4b12-b581-3362d46de89e\" (UID: \"cae75d4f-b139-4b12-b581-3362d46de89e\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.515315 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c97044-4916-4bdd-aae7-33069d43663e-operator-scripts\") pod \"a1c97044-4916-4bdd-aae7-33069d43663e\" (UID: \"a1c97044-4916-4bdd-aae7-33069d43663e\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.515379 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glppg\" (UniqueName: \"kubernetes.io/projected/908b60c2-798e-4feb-90da-4cd08d479578-kube-api-access-glppg\") pod \"908b60c2-798e-4feb-90da-4cd08d479578\" (UID: \"908b60c2-798e-4feb-90da-4cd08d479578\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.515465 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fc8a762-b8f7-4246-aaec-d6670cf715a5-operator-scripts\") pod \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\" (UID: \"2fc8a762-b8f7-4246-aaec-d6670cf715a5\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.515529 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/908b60c2-798e-4feb-90da-4cd08d479578-operator-scripts\") pod \"908b60c2-798e-4feb-90da-4cd08d479578\" (UID: \"908b60c2-798e-4feb-90da-4cd08d479578\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.515554 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ddqt\" (UniqueName: \"kubernetes.io/projected/32b5e812-c653-4437-ae27-b7a1635085a6-kube-api-access-6ddqt\") pod \"32b5e812-c653-4437-ae27-b7a1635085a6\" (UID: \"32b5e812-c653-4437-ae27-b7a1635085a6\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.515601 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b5e812-c653-4437-ae27-b7a1635085a6-operator-scripts\") pod \"32b5e812-c653-4437-ae27-b7a1635085a6\" (UID: \"32b5e812-c653-4437-ae27-b7a1635085a6\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.516262 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/908b60c2-798e-4feb-90da-4cd08d479578-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "908b60c2-798e-4feb-90da-4cd08d479578" (UID: "908b60c2-798e-4feb-90da-4cd08d479578"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.517024 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1c97044-4916-4bdd-aae7-33069d43663e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a1c97044-4916-4bdd-aae7-33069d43663e" (UID: "a1c97044-4916-4bdd-aae7-33069d43663e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.517137 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32b5e812-c653-4437-ae27-b7a1635085a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32b5e812-c653-4437-ae27-b7a1635085a6" (UID: "32b5e812-c653-4437-ae27-b7a1635085a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.517301 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1c97044-4916-4bdd-aae7-33069d43663e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.517315 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/908b60c2-798e-4feb-90da-4cd08d479578-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.517324 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b5e812-c653-4437-ae27-b7a1635085a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.517502 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc8a762-b8f7-4246-aaec-d6670cf715a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2fc8a762-b8f7-4246-aaec-d6670cf715a5" (UID: "2fc8a762-b8f7-4246-aaec-d6670cf715a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.518681 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.520160 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cae75d4f-b139-4b12-b581-3362d46de89e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cae75d4f-b139-4b12-b581-3362d46de89e" (UID: "cae75d4f-b139-4b12-b581-3362d46de89e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.521517 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c97044-4916-4bdd-aae7-33069d43663e-kube-api-access-rmt9r" (OuterVolumeSpecName: "kube-api-access-rmt9r") pod "a1c97044-4916-4bdd-aae7-33069d43663e" (UID: "a1c97044-4916-4bdd-aae7-33069d43663e"). InnerVolumeSpecName "kube-api-access-rmt9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.521578 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32b5e812-c653-4437-ae27-b7a1635085a6-kube-api-access-6ddqt" (OuterVolumeSpecName: "kube-api-access-6ddqt") pod "32b5e812-c653-4437-ae27-b7a1635085a6" (UID: "32b5e812-c653-4437-ae27-b7a1635085a6"). InnerVolumeSpecName "kube-api-access-6ddqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.524656 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fc8a762-b8f7-4246-aaec-d6670cf715a5-kube-api-access-rnxjz" (OuterVolumeSpecName: "kube-api-access-rnxjz") pod "2fc8a762-b8f7-4246-aaec-d6670cf715a5" (UID: "2fc8a762-b8f7-4246-aaec-d6670cf715a5"). InnerVolumeSpecName "kube-api-access-rnxjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.525673 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/908b60c2-798e-4feb-90da-4cd08d479578-kube-api-access-glppg" (OuterVolumeSpecName: "kube-api-access-glppg") pod "908b60c2-798e-4feb-90da-4cd08d479578" (UID: "908b60c2-798e-4feb-90da-4cd08d479578"). InnerVolumeSpecName "kube-api-access-glppg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.530513 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cae75d4f-b139-4b12-b581-3362d46de89e-kube-api-access-vfrnx" (OuterVolumeSpecName: "kube-api-access-vfrnx") pod "cae75d4f-b139-4b12-b581-3362d46de89e" (UID: "cae75d4f-b139-4b12-b581-3362d46de89e"). InnerVolumeSpecName "kube-api-access-vfrnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618453 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swh9m\" (UniqueName: \"kubernetes.io/projected/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-kube-api-access-swh9m\") pod \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618491 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-config\") pod \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618540 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc04d3ff-c5b1-4134-b544-ca9341af72b4-operator-scripts\") pod \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\" (UID: \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618569 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pnlx\" (UniqueName: \"kubernetes.io/projected/728c9997-66cf-4996-8b4d-75af0e3ba55c-kube-api-access-7pnlx\") pod \"728c9997-66cf-4996-8b4d-75af0e3ba55c\" (UID: \"728c9997-66cf-4996-8b4d-75af0e3ba55c\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618701 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-sb\") pod \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618718 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5254b05c-78e8-49fa-932b-4f36b5fb907d-operator-scripts\") pod \"5254b05c-78e8-49fa-932b-4f36b5fb907d\" (UID: \"5254b05c-78e8-49fa-932b-4f36b5fb907d\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618746 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-dns-svc\") pod \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618779 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8rsp\" (UniqueName: \"kubernetes.io/projected/5254b05c-78e8-49fa-932b-4f36b5fb907d-kube-api-access-s8rsp\") pod \"5254b05c-78e8-49fa-932b-4f36b5fb907d\" (UID: \"5254b05c-78e8-49fa-932b-4f36b5fb907d\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618808 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-nb\") pod \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\" (UID: \"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618840 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdk89\" (UniqueName: \"kubernetes.io/projected/bc04d3ff-c5b1-4134-b544-ca9341af72b4-kube-api-access-hdk89\") pod \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\" (UID: \"bc04d3ff-c5b1-4134-b544-ca9341af72b4\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.618878 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/728c9997-66cf-4996-8b4d-75af0e3ba55c-operator-scripts\") pod \"728c9997-66cf-4996-8b4d-75af0e3ba55c\" (UID: \"728c9997-66cf-4996-8b4d-75af0e3ba55c\") " Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.619356 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnxjz\" (UniqueName: \"kubernetes.io/projected/2fc8a762-b8f7-4246-aaec-d6670cf715a5-kube-api-access-rnxjz\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.619382 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmt9r\" (UniqueName: \"kubernetes.io/projected/a1c97044-4916-4bdd-aae7-33069d43663e-kube-api-access-rmt9r\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.619393 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae75d4f-b139-4b12-b581-3362d46de89e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.619403 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfrnx\" (UniqueName: \"kubernetes.io/projected/cae75d4f-b139-4b12-b581-3362d46de89e-kube-api-access-vfrnx\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.619412 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glppg\" (UniqueName: \"kubernetes.io/projected/908b60c2-798e-4feb-90da-4cd08d479578-kube-api-access-glppg\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.619422 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fc8a762-b8f7-4246-aaec-d6670cf715a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.619431 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ddqt\" (UniqueName: \"kubernetes.io/projected/32b5e812-c653-4437-ae27-b7a1635085a6-kube-api-access-6ddqt\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.619761 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/728c9997-66cf-4996-8b4d-75af0e3ba55c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "728c9997-66cf-4996-8b4d-75af0e3ba55c" (UID: "728c9997-66cf-4996-8b4d-75af0e3ba55c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.620096 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5254b05c-78e8-49fa-932b-4f36b5fb907d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5254b05c-78e8-49fa-932b-4f36b5fb907d" (UID: "5254b05c-78e8-49fa-932b-4f36b5fb907d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.621446 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-kube-api-access-swh9m" (OuterVolumeSpecName: "kube-api-access-swh9m") pod "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" (UID: "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176"). InnerVolumeSpecName "kube-api-access-swh9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.621465 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc04d3ff-c5b1-4134-b544-ca9341af72b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc04d3ff-c5b1-4134-b544-ca9341af72b4" (UID: "bc04d3ff-c5b1-4134-b544-ca9341af72b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.624241 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5254b05c-78e8-49fa-932b-4f36b5fb907d-kube-api-access-s8rsp" (OuterVolumeSpecName: "kube-api-access-s8rsp") pod "5254b05c-78e8-49fa-932b-4f36b5fb907d" (UID: "5254b05c-78e8-49fa-932b-4f36b5fb907d"). InnerVolumeSpecName "kube-api-access-s8rsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.624712 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc04d3ff-c5b1-4134-b544-ca9341af72b4-kube-api-access-hdk89" (OuterVolumeSpecName: "kube-api-access-hdk89") pod "bc04d3ff-c5b1-4134-b544-ca9341af72b4" (UID: "bc04d3ff-c5b1-4134-b544-ca9341af72b4"). InnerVolumeSpecName "kube-api-access-hdk89". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.624780 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/728c9997-66cf-4996-8b4d-75af0e3ba55c-kube-api-access-7pnlx" (OuterVolumeSpecName: "kube-api-access-7pnlx") pod "728c9997-66cf-4996-8b4d-75af0e3ba55c" (UID: "728c9997-66cf-4996-8b4d-75af0e3ba55c"). InnerVolumeSpecName "kube-api-access-7pnlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.660809 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-config" (OuterVolumeSpecName: "config") pod "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" (UID: "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.665381 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" (UID: "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.670022 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" (UID: "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.673994 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" (UID: "ed52e50d-2db0-4bbe-93a8-ae4d00f6f176"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721193 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/728c9997-66cf-4996-8b4d-75af0e3ba55c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721234 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swh9m\" (UniqueName: \"kubernetes.io/projected/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-kube-api-access-swh9m\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721276 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721285 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc04d3ff-c5b1-4134-b544-ca9341af72b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721294 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pnlx\" (UniqueName: \"kubernetes.io/projected/728c9997-66cf-4996-8b4d-75af0e3ba55c-kube-api-access-7pnlx\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721302 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721310 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5254b05c-78e8-49fa-932b-4f36b5fb907d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721318 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721328 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8rsp\" (UniqueName: \"kubernetes.io/projected/5254b05c-78e8-49fa-932b-4f36b5fb907d-kube-api-access-s8rsp\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721337 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.721366 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdk89\" (UniqueName: \"kubernetes.io/projected/bc04d3ff-c5b1-4134-b544-ca9341af72b4-kube-api-access-hdk89\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.762476 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bhnlm" event={"ID":"908b60c2-798e-4feb-90da-4cd08d479578","Type":"ContainerDied","Data":"38eeb7779e82002497451f647f3c2d893f2cb8fe4e809c842e808c4321adb6f0"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.762525 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38eeb7779e82002497451f647f3c2d893f2cb8fe4e809c842e808c4321adb6f0" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.762594 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bhnlm" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.768712 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-83e5-account-create-ttzgw" event={"ID":"32b5e812-c653-4437-ae27-b7a1635085a6","Type":"ContainerDied","Data":"b7fa3e4a4b44bf175fa8057abc34dfa722a82914ce546d01bd12494dd728fe94"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.768749 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7fa3e4a4b44bf175fa8057abc34dfa722a82914ce546d01bd12494dd728fe94" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.768802 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-83e5-account-create-ttzgw" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.775586 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-ltdt2" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.775585 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-ltdt2" event={"ID":"5254b05c-78e8-49fa-932b-4f36b5fb907d","Type":"ContainerDied","Data":"1ad573ed4dd23c5f53ed7fcb297baf46ad76fc0b257cd38e8feba6e50d037ad4"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.776259 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ad573ed4dd23c5f53ed7fcb297baf46ad76fc0b257cd38e8feba6e50d037ad4" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.790106 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-s9xbh" event={"ID":"ed52e50d-2db0-4bbe-93a8-ae4d00f6f176","Type":"ContainerDied","Data":"7f0bd4c931b4031a77f2db9c356ba6fe0fc79de1c8093280e06ec20018a580d5"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.790178 4822 scope.go:117] "RemoveContainer" containerID="5df2a9dad76627891e3173a6132be88f4f251a88fcb642e77f4cfb5d40008de3" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.790402 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-s9xbh" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.793053 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-57687" event={"ID":"274d41ac-17fe-48ef-ad92-f769c96b219b","Type":"ContainerStarted","Data":"91b1804b1e534f9081e6b584352489cfb8d984eef5a0764bc62dffe333e3bacd"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.794652 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4kdxz" event={"ID":"bc04d3ff-c5b1-4134-b544-ca9341af72b4","Type":"ContainerDied","Data":"df2bae0321c9376a4d5408425737723f4478555abad84bf2c3aea731e143f2a1"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.794720 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df2bae0321c9376a4d5408425737723f4478555abad84bf2c3aea731e143f2a1" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.794658 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4kdxz" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.795921 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lhp86" event={"ID":"cae75d4f-b139-4b12-b581-3362d46de89e","Type":"ContainerDied","Data":"a4778d288dca17ad1d3de324e73393ab2905d4dd4ffcedf43ff1c5e3ef82a4bb"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.795944 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4778d288dca17ad1d3de324e73393ab2905d4dd4ffcedf43ff1c5e3ef82a4bb" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.795982 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lhp86" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.797625 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-3aed-account-create-db4jq" event={"ID":"a1c97044-4916-4bdd-aae7-33069d43663e","Type":"ContainerDied","Data":"c2097bd114bfbddbaf73a4eb52f9e30a14f309655d1863a081957de6a1dc26eb"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.797645 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2097bd114bfbddbaf73a4eb52f9e30a14f309655d1863a081957de6a1dc26eb" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.797679 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-3aed-account-create-db4jq" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.807355 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8fc4-account-create-tgmfq" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.807754 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8fc4-account-create-tgmfq" event={"ID":"2fc8a762-b8f7-4246-aaec-d6670cf715a5","Type":"ContainerDied","Data":"24b41a29acc9a11d2e68a4be013da9cb89289950e3e9f34d6a070641630772c7"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.807946 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24b41a29acc9a11d2e68a4be013da9cb89289950e3e9f34d6a070641630772c7" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.812407 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5f21-account-create-ds2h6" event={"ID":"728c9997-66cf-4996-8b4d-75af0e3ba55c","Type":"ContainerDied","Data":"fa30fe893432cdf707216cba5f48ceb8041682db66d3494d4c1a0526e81e5585"} Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.812579 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa30fe893432cdf707216cba5f48ceb8041682db66d3494d4c1a0526e81e5585" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.812402 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5f21-account-create-ds2h6" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.824694 4822 scope.go:117] "RemoveContainer" containerID="f97bdf86e5b2c340231c05300811df6caf732375c3943c2abf5b0cf33f8a074c" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.826282 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-57687" podStartSLOduration=2.226391878 podStartE2EDuration="7.826271686s" podCreationTimestamp="2025-11-24 14:37:21 +0000 UTC" firstStartedPulling="2025-11-24 14:37:22.561551273 +0000 UTC m=+1079.678191740" lastFinishedPulling="2025-11-24 14:37:28.161431071 +0000 UTC m=+1085.278071548" observedRunningTime="2025-11-24 14:37:28.822352092 +0000 UTC m=+1085.938992579" watchObservedRunningTime="2025-11-24 14:37:28.826271686 +0000 UTC m=+1085.942912163" Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.881897 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-s9xbh"] Nov 24 14:37:28 crc kubenswrapper[4822]: I1124 14:37:28.890263 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-s9xbh"] Nov 24 14:37:29 crc kubenswrapper[4822]: I1124 14:37:29.719505 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" path="/var/lib/kubelet/pods/ed52e50d-2db0-4bbe-93a8-ae4d00f6f176/volumes" Nov 24 14:37:31 crc kubenswrapper[4822]: I1124 14:37:31.860012 4822 generic.go:334] "Generic (PLEG): container finished" podID="274d41ac-17fe-48ef-ad92-f769c96b219b" containerID="91b1804b1e534f9081e6b584352489cfb8d984eef5a0764bc62dffe333e3bacd" exitCode=0 Nov 24 14:37:31 crc kubenswrapper[4822]: I1124 14:37:31.860095 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-57687" event={"ID":"274d41ac-17fe-48ef-ad92-f769c96b219b","Type":"ContainerDied","Data":"91b1804b1e534f9081e6b584352489cfb8d984eef5a0764bc62dffe333e3bacd"} Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.393730 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-57687" Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.520470 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc7cn\" (UniqueName: \"kubernetes.io/projected/274d41ac-17fe-48ef-ad92-f769c96b219b-kube-api-access-rc7cn\") pod \"274d41ac-17fe-48ef-ad92-f769c96b219b\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.520817 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-config-data\") pod \"274d41ac-17fe-48ef-ad92-f769c96b219b\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.520863 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-combined-ca-bundle\") pod \"274d41ac-17fe-48ef-ad92-f769c96b219b\" (UID: \"274d41ac-17fe-48ef-ad92-f769c96b219b\") " Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.529525 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/274d41ac-17fe-48ef-ad92-f769c96b219b-kube-api-access-rc7cn" (OuterVolumeSpecName: "kube-api-access-rc7cn") pod "274d41ac-17fe-48ef-ad92-f769c96b219b" (UID: "274d41ac-17fe-48ef-ad92-f769c96b219b"). InnerVolumeSpecName "kube-api-access-rc7cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.552697 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "274d41ac-17fe-48ef-ad92-f769c96b219b" (UID: "274d41ac-17fe-48ef-ad92-f769c96b219b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.574673 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-config-data" (OuterVolumeSpecName: "config-data") pod "274d41ac-17fe-48ef-ad92-f769c96b219b" (UID: "274d41ac-17fe-48ef-ad92-f769c96b219b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.623723 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc7cn\" (UniqueName: \"kubernetes.io/projected/274d41ac-17fe-48ef-ad92-f769c96b219b-kube-api-access-rc7cn\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.623778 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.623798 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274d41ac-17fe-48ef-ad92-f769c96b219b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.884859 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-57687" event={"ID":"274d41ac-17fe-48ef-ad92-f769c96b219b","Type":"ContainerDied","Data":"7a05ae2f1e523c8a96f5f349a2166f0d6c795dbf1dff11f9c88137d4b27fcbb2"} Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.884901 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a05ae2f1e523c8a96f5f349a2166f0d6c795dbf1dff11f9c88137d4b27fcbb2" Nov 24 14:37:33 crc kubenswrapper[4822]: I1124 14:37:33.885009 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-57687" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299238 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hbspf"] Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299562 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" containerName="init" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299593 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" containerName="init" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299607 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c97044-4916-4bdd-aae7-33069d43663e" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299613 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c97044-4916-4bdd-aae7-33069d43663e" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299627 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae75d4f-b139-4b12-b581-3362d46de89e" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299633 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae75d4f-b139-4b12-b581-3362d46de89e" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299646 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" containerName="dnsmasq-dns" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299652 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" containerName="dnsmasq-dns" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299665 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274d41ac-17fe-48ef-ad92-f769c96b219b" containerName="keystone-db-sync" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299671 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="274d41ac-17fe-48ef-ad92-f769c96b219b" containerName="keystone-db-sync" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299686 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc8a762-b8f7-4246-aaec-d6670cf715a5" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299782 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc8a762-b8f7-4246-aaec-d6670cf715a5" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299807 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5254b05c-78e8-49fa-932b-4f36b5fb907d" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299814 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5254b05c-78e8-49fa-932b-4f36b5fb907d" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299823 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b5e812-c653-4437-ae27-b7a1635085a6" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299829 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b5e812-c653-4437-ae27-b7a1635085a6" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299840 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="728c9997-66cf-4996-8b4d-75af0e3ba55c" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299847 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="728c9997-66cf-4996-8b4d-75af0e3ba55c" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299856 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="908b60c2-798e-4feb-90da-4cd08d479578" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299862 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="908b60c2-798e-4feb-90da-4cd08d479578" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: E1124 14:37:34.299875 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc04d3ff-c5b1-4134-b544-ca9341af72b4" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.299882 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc04d3ff-c5b1-4134-b544-ca9341af72b4" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300253 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="274d41ac-17fe-48ef-ad92-f769c96b219b" containerName="keystone-db-sync" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300268 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc04d3ff-c5b1-4134-b544-ca9341af72b4" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300280 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="908b60c2-798e-4feb-90da-4cd08d479578" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300293 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae75d4f-b139-4b12-b581-3362d46de89e" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300303 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="32b5e812-c653-4437-ae27-b7a1635085a6" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300311 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c97044-4916-4bdd-aae7-33069d43663e" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300321 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="5254b05c-78e8-49fa-932b-4f36b5fb907d" containerName="mariadb-database-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300333 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed52e50d-2db0-4bbe-93a8-ae4d00f6f176" containerName="dnsmasq-dns" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300341 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="728c9997-66cf-4996-8b4d-75af0e3ba55c" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300353 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fc8a762-b8f7-4246-aaec-d6670cf715a5" containerName="mariadb-account-create" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.300979 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.308844 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.309027 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r2n9r" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.309081 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.309047 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.309283 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.316033 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-784f69c749-hj58z"] Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.318373 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.343134 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hbspf"] Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.367295 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-hj58z"] Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452339 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-credential-keys\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452424 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452468 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-config-data\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452516 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-dns-svc\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452549 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-config\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452705 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-scripts\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452744 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452767 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-combined-ca-bundle\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452781 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbshj\" (UniqueName: \"kubernetes.io/projected/9498898a-d1db-4c10-a06b-f2e0af5f8360-kube-api-access-rbshj\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452817 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nft6\" (UniqueName: \"kubernetes.io/projected/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-kube-api-access-9nft6\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.452844 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-fernet-keys\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555139 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-combined-ca-bundle\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555471 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbshj\" (UniqueName: \"kubernetes.io/projected/9498898a-d1db-4c10-a06b-f2e0af5f8360-kube-api-access-rbshj\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555493 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nft6\" (UniqueName: \"kubernetes.io/projected/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-kube-api-access-9nft6\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555520 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-fernet-keys\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555544 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-credential-keys\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555582 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555621 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-config-data\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555647 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-dns-svc\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555678 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-config\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555740 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-scripts\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.555759 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.556594 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.561993 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-config\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.562542 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-dns-svc\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.563044 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.569000 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-combined-ca-bundle\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.576252 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-fernet-keys\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.579845 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-credential-keys\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.580263 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-config-data\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.580979 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-scripts\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.593869 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-fzrq8"] Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.597981 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbshj\" (UniqueName: \"kubernetes.io/projected/9498898a-d1db-4c10-a06b-f2e0af5f8360-kube-api-access-rbshj\") pod \"dnsmasq-dns-784f69c749-hj58z\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.599173 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.641725 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-fzrq8"] Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.642631 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-fzfh7" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.642836 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.642975 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.653453 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nft6\" (UniqueName: \"kubernetes.io/projected/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-kube-api-access-9nft6\") pod \"keystone-bootstrap-hbspf\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.697875 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-8rtpz"] Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.707011 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.710932 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.711127 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2mjb8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.711374 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.718057 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.719182 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-config\") pod \"neutron-db-sync-8rtpz\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.719305 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjf24\" (UniqueName: \"kubernetes.io/projected/35fdf286-3217-480f-9522-3d146be7ceab-kube-api-access-fjf24\") pod \"neutron-db-sync-8rtpz\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.719338 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-combined-ca-bundle\") pod \"neutron-db-sync-8rtpz\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.797283 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-8rtpz"] Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.843767 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjf24\" (UniqueName: \"kubernetes.io/projected/35fdf286-3217-480f-9522-3d146be7ceab-kube-api-access-fjf24\") pod \"neutron-db-sync-8rtpz\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.843813 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r72z7\" (UniqueName: \"kubernetes.io/projected/7348b48b-1f84-439f-a049-bfc19935dbfc-kube-api-access-r72z7\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.843843 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-combined-ca-bundle\") pod \"neutron-db-sync-8rtpz\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.843895 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7348b48b-1f84-439f-a049-bfc19935dbfc-etc-machine-id\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.843913 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-config-data\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.843927 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-combined-ca-bundle\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.843974 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-config\") pod \"neutron-db-sync-8rtpz\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.844002 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-scripts\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.844041 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-db-sync-config-data\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.854728 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-config\") pod \"neutron-db-sync-8rtpz\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.865611 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-combined-ca-bundle\") pod \"neutron-db-sync-8rtpz\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.877441 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjf24\" (UniqueName: \"kubernetes.io/projected/35fdf286-3217-480f-9522-3d146be7ceab-kube-api-access-fjf24\") pod \"neutron-db-sync-8rtpz\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.890450 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.941598 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.948655 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-scripts\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.948712 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-db-sync-config-data\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.948764 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r72z7\" (UniqueName: \"kubernetes.io/projected/7348b48b-1f84-439f-a049-bfc19935dbfc-kube-api-access-r72z7\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.948821 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7348b48b-1f84-439f-a049-bfc19935dbfc-etc-machine-id\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.948840 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-config-data\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.948854 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-combined-ca-bundle\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.950967 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7348b48b-1f84-439f-a049-bfc19935dbfc-etc-machine-id\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.962076 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-config-data\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.962410 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-scripts\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.963607 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-combined-ca-bundle\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.966751 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-db-sync-config-data\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.972720 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r72z7\" (UniqueName: \"kubernetes.io/projected/7348b48b-1f84-439f-a049-bfc19935dbfc-kube-api-access-r72z7\") pod \"cinder-db-sync-fzrq8\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.986801 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-fxf96"] Nov 24 14:37:34 crc kubenswrapper[4822]: I1124 14:37:34.988197 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.002379 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-bdtcv" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.002646 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.013479 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-fxf96"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.027671 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-hj58z"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.060939 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-s6qvf"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.062273 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.065363 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.066504 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-zhp7h" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.068065 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.070977 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-s6qvf"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.072347 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.070388 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.093101 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-227wq"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.094549 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.099881 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rv2r2" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.100046 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.100497 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.122432 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gmz62"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.128534 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.129768 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-227wq"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.146566 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.148929 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.151567 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.151633 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.151748 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-db-sync-config-data\") pod \"barbican-db-sync-fxf96\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.151833 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-combined-ca-bundle\") pod \"barbican-db-sync-fxf96\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.152861 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-config-data\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.152898 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-scripts\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.152953 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnz6m\" (UniqueName: \"kubernetes.io/projected/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-kube-api-access-gnz6m\") pod \"barbican-db-sync-fxf96\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.153011 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-certs\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.153043 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-combined-ca-bundle\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.153752 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx97z\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-kube-api-access-rx97z\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.166504 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gmz62"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.178003 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257188 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnz6m\" (UniqueName: \"kubernetes.io/projected/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-kube-api-access-gnz6m\") pod \"barbican-db-sync-fxf96\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257245 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257277 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-config\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257303 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-scripts\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257381 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-certs\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257413 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-combined-ca-bundle\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257438 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx97z\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-kube-api-access-rx97z\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257463 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-scripts\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257479 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k9qd\" (UniqueName: \"kubernetes.io/projected/d2cfadc6-5078-4554-849b-c013b8fa0a2e-kube-api-access-6k9qd\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257501 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2khl7\" (UniqueName: \"kubernetes.io/projected/1ae8141e-86cb-43bc-9ada-307337b1566f-kube-api-access-2khl7\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257520 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-config-data\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257539 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-combined-ca-bundle\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257574 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-db-sync-config-data\") pod \"barbican-db-sync-fxf96\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257605 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-log-httpd\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257622 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-config-data\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257640 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257657 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257689 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-run-httpd\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257714 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-dns-svc\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257739 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257762 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-combined-ca-bundle\") pod \"barbican-db-sync-fxf96\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257780 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz25l\" (UniqueName: \"kubernetes.io/projected/42faa16a-d5df-4002-94bc-dd9b727ad202-kube-api-access-dz25l\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257797 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-config-data\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257821 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-scripts\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.257852 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42faa16a-d5df-4002-94bc-dd9b727ad202-logs\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.263093 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-combined-ca-bundle\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.263271 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-db-sync-config-data\") pod \"barbican-db-sync-fxf96\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.263863 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-combined-ca-bundle\") pod \"barbican-db-sync-fxf96\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.265567 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-config-data\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.265948 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-scripts\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.266793 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-certs\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.273939 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnz6m\" (UniqueName: \"kubernetes.io/projected/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-kube-api-access-gnz6m\") pod \"barbican-db-sync-fxf96\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.286883 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx97z\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-kube-api-access-rx97z\") pod \"cloudkitty-db-sync-s6qvf\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.323147 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fxf96" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359373 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-scripts\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359419 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k9qd\" (UniqueName: \"kubernetes.io/projected/d2cfadc6-5078-4554-849b-c013b8fa0a2e-kube-api-access-6k9qd\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359445 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2khl7\" (UniqueName: \"kubernetes.io/projected/1ae8141e-86cb-43bc-9ada-307337b1566f-kube-api-access-2khl7\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359465 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-config-data\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359483 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-combined-ca-bundle\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359519 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-log-httpd\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359539 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-config-data\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359558 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359576 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359601 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-run-httpd\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359622 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-dns-svc\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359643 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359666 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz25l\" (UniqueName: \"kubernetes.io/projected/42faa16a-d5df-4002-94bc-dd9b727ad202-kube-api-access-dz25l\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359709 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42faa16a-d5df-4002-94bc-dd9b727ad202-logs\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359746 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359768 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-config\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.359784 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-scripts\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.360940 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.361790 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-log-httpd\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.362464 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-run-httpd\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.363040 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-scripts\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.363269 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-dns-svc\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.363789 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.364023 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42faa16a-d5df-4002-94bc-dd9b727ad202-logs\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.364545 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-config\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.364779 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-scripts\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.366622 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.366830 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-config-data\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.369789 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.372077 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-config-data\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.374066 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-combined-ca-bundle\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.378081 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k9qd\" (UniqueName: \"kubernetes.io/projected/d2cfadc6-5078-4554-849b-c013b8fa0a2e-kube-api-access-6k9qd\") pod \"dnsmasq-dns-f84976bdf-gmz62\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.378173 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2khl7\" (UniqueName: \"kubernetes.io/projected/1ae8141e-86cb-43bc-9ada-307337b1566f-kube-api-access-2khl7\") pod \"ceilometer-0\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.378478 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz25l\" (UniqueName: \"kubernetes.io/projected/42faa16a-d5df-4002-94bc-dd9b727ad202-kube-api-access-dz25l\") pod \"placement-db-sync-227wq\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.463653 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.485842 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-227wq" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.523825 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.524685 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.529701 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.531070 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-l2kgh" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.532442 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.532671 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.533782 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.543708 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.548235 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.573406 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-8rtpz"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.623952 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.632054 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.635833 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.638941 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.641523 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.666659 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.666735 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-config-data\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.667212 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-logs\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.667240 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-scripts\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.667266 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.667319 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw6j8\" (UniqueName: \"kubernetes.io/projected/00a35b04-c993-40b5-8da8-8f19ce23a03b-kube-api-access-rw6j8\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.667349 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.667528 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.689948 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-hj58z"] Nov 24 14:37:35 crc kubenswrapper[4822]: W1124 14:37:35.717554 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9498898a_d1db_4c10_a06b_f2e0af5f8360.slice/crio-a015b04aabe168cd0cc43a1d00178768e7e74a3188c97b0d11bb6d5a9525e62f WatchSource:0}: Error finding container a015b04aabe168cd0cc43a1d00178768e7e74a3188c97b0d11bb6d5a9525e62f: Status 404 returned error can't find the container with id a015b04aabe168cd0cc43a1d00178768e7e74a3188c97b0d11bb6d5a9525e62f Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.770548 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw6j8\" (UniqueName: \"kubernetes.io/projected/00a35b04-c993-40b5-8da8-8f19ce23a03b-kube-api-access-rw6j8\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.770803 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.770845 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.770875 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-logs\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.770963 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771049 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771120 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771216 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-config-data\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771300 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771341 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771367 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-logs\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771394 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-scripts\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771419 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfwxc\" (UniqueName: \"kubernetes.io/projected/caf6e332-337b-4356-9bdf-d3ddf3d92e59-kube-api-access-hfwxc\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771454 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771486 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-scripts\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771513 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-config-data\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.771999 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.773796 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-fzrq8"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.781117 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-logs\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.782410 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.783090 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.783179 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2ffde4dda1be13a8596eb6948dbe2a502fe570389c8cc35530979ae06145ae21/globalmount\"" pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.787250 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-config-data\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.787507 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.795384 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-scripts\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.795603 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hbspf"] Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.804114 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw6j8\" (UniqueName: \"kubernetes.io/projected/00a35b04-c993-40b5-8da8-8f19ce23a03b-kube-api-access-rw6j8\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: W1124 14:37:35.835263 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1726f74b_ca5e_43ac_9b0e_752ff7d63a9f.slice/crio-81a4b0fcde9662886d8ecb66371e53becb64aef7b74978d0ebdaec2f72dde24f WatchSource:0}: Error finding container 81a4b0fcde9662886d8ecb66371e53becb64aef7b74978d0ebdaec2f72dde24f: Status 404 returned error can't find the container with id 81a4b0fcde9662886d8ecb66371e53becb64aef7b74978d0ebdaec2f72dde24f Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.879243 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.879336 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.879361 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.879383 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfwxc\" (UniqueName: \"kubernetes.io/projected/caf6e332-337b-4356-9bdf-d3ddf3d92e59-kube-api-access-hfwxc\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.879597 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-scripts\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.879620 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-config-data\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.879715 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.879735 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-logs\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.879820 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.881022 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.882975 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-scripts\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.883390 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-logs\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.887246 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.887284 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4ad6c7c6bc95d7d12c286d711c7d508fc5826003cdc7d089436fe350e33badd7/globalmount\"" pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.890308 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.890527 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.891400 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-config-data\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.896593 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfwxc\" (UniqueName: \"kubernetes.io/projected/caf6e332-337b-4356-9bdf-d3ddf3d92e59-kube-api-access-hfwxc\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.931446 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fzrq8" event={"ID":"7348b48b-1f84-439f-a049-bfc19935dbfc","Type":"ContainerStarted","Data":"b496b8bc17235a62e228ad291c7b1ab2a3bcdaead4c0a12b56811dae29e13346"} Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.932465 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hbspf" event={"ID":"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f","Type":"ContainerStarted","Data":"81a4b0fcde9662886d8ecb66371e53becb64aef7b74978d0ebdaec2f72dde24f"} Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.934532 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8rtpz" event={"ID":"35fdf286-3217-480f-9522-3d146be7ceab","Type":"ContainerStarted","Data":"dbc79fef58ca58767d123fd6668334e7824fecd708b4a8ea53ccf4f21559adab"} Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.938992 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-hj58z" event={"ID":"9498898a-d1db-4c10-a06b-f2e0af5f8360","Type":"ContainerStarted","Data":"a015b04aabe168cd0cc43a1d00178768e7e74a3188c97b0d11bb6d5a9525e62f"} Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.966621 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:35 crc kubenswrapper[4822]: I1124 14:37:35.971947 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-fxf96"] Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.084760 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-s6qvf"] Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.152604 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.159963 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.201624 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gmz62"] Nov 24 14:37:36 crc kubenswrapper[4822]: W1124 14:37:36.219648 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42faa16a_d5df_4002_94bc_dd9b727ad202.slice/crio-e42521fcb2ef0acae9391961e55ec0f9aef8f204a0043b83ab411828b5c088ee WatchSource:0}: Error finding container e42521fcb2ef0acae9391961e55ec0f9aef8f204a0043b83ab411828b5c088ee: Status 404 returned error can't find the container with id e42521fcb2ef0acae9391961e55ec0f9aef8f204a0043b83ab411828b5c088ee Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.229675 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-227wq"] Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.277822 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.806647 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.906622 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:37:36 crc kubenswrapper[4822]: W1124 14:37:36.919675 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaf6e332_337b_4356_9bdf_d3ddf3d92e59.slice/crio-6a60118fcc880377c9a9dcc99f6e20baeefde3907de3742182f06770d9f406db WatchSource:0}: Error finding container 6a60118fcc880377c9a9dcc99f6e20baeefde3907de3742182f06770d9f406db: Status 404 returned error can't find the container with id 6a60118fcc880377c9a9dcc99f6e20baeefde3907de3742182f06770d9f406db Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.973512 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a35b04-c993-40b5-8da8-8f19ce23a03b","Type":"ContainerStarted","Data":"ec7e886af5719edd61ed865f9a3a1b69fac243c54e7fddf5e7e4b3c52516f3ee"} Nov 24 14:37:36 crc kubenswrapper[4822]: I1124 14:37:36.998363 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-227wq" event={"ID":"42faa16a-d5df-4002-94bc-dd9b727ad202","Type":"ContainerStarted","Data":"e42521fcb2ef0acae9391961e55ec0f9aef8f204a0043b83ab411828b5c088ee"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.001539 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fxf96" event={"ID":"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd","Type":"ContainerStarted","Data":"9d5b49489757cce5841d9b4c6364c2f74eada35a8068544cca3572dd309628e3"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.004498 4822 generic.go:334] "Generic (PLEG): container finished" podID="d2cfadc6-5078-4554-849b-c013b8fa0a2e" containerID="4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9" exitCode=0 Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.004613 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" event={"ID":"d2cfadc6-5078-4554-849b-c013b8fa0a2e","Type":"ContainerDied","Data":"4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.004659 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" event={"ID":"d2cfadc6-5078-4554-849b-c013b8fa0a2e","Type":"ContainerStarted","Data":"667266956e195cff7e79d05e9e920e4591b7f186a798de69a37eeda0c649a2fc"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.015231 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"caf6e332-337b-4356-9bdf-d3ddf3d92e59","Type":"ContainerStarted","Data":"6a60118fcc880377c9a9dcc99f6e20baeefde3907de3742182f06770d9f406db"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.030772 4822 generic.go:334] "Generic (PLEG): container finished" podID="9498898a-d1db-4c10-a06b-f2e0af5f8360" containerID="b153e2957435bffbf591cf33e82e6ca36cb25843f38f81ff572770fd522f39e4" exitCode=0 Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.030840 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-hj58z" event={"ID":"9498898a-d1db-4c10-a06b-f2e0af5f8360","Type":"ContainerDied","Data":"b153e2957435bffbf591cf33e82e6ca36cb25843f38f81ff572770fd522f39e4"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.037820 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hbspf" event={"ID":"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f","Type":"ContainerStarted","Data":"478c0129945058472624762055c6cbbcff7171196e7b8e7b05755afd0d7ab513"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.044477 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8rtpz" event={"ID":"35fdf286-3217-480f-9522-3d146be7ceab","Type":"ContainerStarted","Data":"5ef5cc9c7238acea58a84c5a18958f784a38b03e541d20ea658682453ba33a37"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.096729 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hbspf" podStartSLOduration=3.096715848 podStartE2EDuration="3.096715848s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:37:37.090229807 +0000 UTC m=+1094.206870284" watchObservedRunningTime="2025-11-24 14:37:37.096715848 +0000 UTC m=+1094.213356325" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.101186 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerStarted","Data":"ba93bc80bc96c1fe1bf2b47001c78988d78b1a1a8aaeac6cedc5cdd15cca7d84"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.123010 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-8rtpz" podStartSLOduration=3.122993598 podStartE2EDuration="3.122993598s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:37:37.117666732 +0000 UTC m=+1094.234307209" watchObservedRunningTime="2025-11-24 14:37:37.122993598 +0000 UTC m=+1094.239634075" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.143239 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-s6qvf" event={"ID":"f0ad3a91-d60c-40a0-a285-c5a9615fffe0","Type":"ContainerStarted","Data":"a6e070c2971b0b872f6258f250bdc1be079cab6033f4aaee409f90bf978a4c9c"} Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.357921 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.498711 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.536603 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.591619 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.640437 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbshj\" (UniqueName: \"kubernetes.io/projected/9498898a-d1db-4c10-a06b-f2e0af5f8360-kube-api-access-rbshj\") pod \"9498898a-d1db-4c10-a06b-f2e0af5f8360\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.640717 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-nb\") pod \"9498898a-d1db-4c10-a06b-f2e0af5f8360\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.640768 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-dns-svc\") pod \"9498898a-d1db-4c10-a06b-f2e0af5f8360\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.640969 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-sb\") pod \"9498898a-d1db-4c10-a06b-f2e0af5f8360\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.641002 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-config\") pod \"9498898a-d1db-4c10-a06b-f2e0af5f8360\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.657515 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9498898a-d1db-4c10-a06b-f2e0af5f8360-kube-api-access-rbshj" (OuterVolumeSpecName: "kube-api-access-rbshj") pod "9498898a-d1db-4c10-a06b-f2e0af5f8360" (UID: "9498898a-d1db-4c10-a06b-f2e0af5f8360"). InnerVolumeSpecName "kube-api-access-rbshj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.674587 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-config" (OuterVolumeSpecName: "config") pod "9498898a-d1db-4c10-a06b-f2e0af5f8360" (UID: "9498898a-d1db-4c10-a06b-f2e0af5f8360"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.699387 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9498898a-d1db-4c10-a06b-f2e0af5f8360" (UID: "9498898a-d1db-4c10-a06b-f2e0af5f8360"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.742221 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9498898a-d1db-4c10-a06b-f2e0af5f8360" (UID: "9498898a-d1db-4c10-a06b-f2e0af5f8360"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.742631 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-dns-svc\") pod \"9498898a-d1db-4c10-a06b-f2e0af5f8360\" (UID: \"9498898a-d1db-4c10-a06b-f2e0af5f8360\") " Nov 24 14:37:37 crc kubenswrapper[4822]: W1124 14:37:37.742851 4822 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/9498898a-d1db-4c10-a06b-f2e0af5f8360/volumes/kubernetes.io~configmap/dns-svc Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.742906 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9498898a-d1db-4c10-a06b-f2e0af5f8360" (UID: "9498898a-d1db-4c10-a06b-f2e0af5f8360"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.743780 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.743795 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.743808 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.743818 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbshj\" (UniqueName: \"kubernetes.io/projected/9498898a-d1db-4c10-a06b-f2e0af5f8360-kube-api-access-rbshj\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.821811 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9498898a-d1db-4c10-a06b-f2e0af5f8360" (UID: "9498898a-d1db-4c10-a06b-f2e0af5f8360"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:37:37 crc kubenswrapper[4822]: I1124 14:37:37.846069 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9498898a-d1db-4c10-a06b-f2e0af5f8360-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:38 crc kubenswrapper[4822]: I1124 14:37:38.168275 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" event={"ID":"d2cfadc6-5078-4554-849b-c013b8fa0a2e","Type":"ContainerStarted","Data":"4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae"} Nov 24 14:37:38 crc kubenswrapper[4822]: I1124 14:37:38.168322 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:38 crc kubenswrapper[4822]: I1124 14:37:38.172302 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-hj58z" Nov 24 14:37:38 crc kubenswrapper[4822]: I1124 14:37:38.173686 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-hj58z" event={"ID":"9498898a-d1db-4c10-a06b-f2e0af5f8360","Type":"ContainerDied","Data":"a015b04aabe168cd0cc43a1d00178768e7e74a3188c97b0d11bb6d5a9525e62f"} Nov 24 14:37:38 crc kubenswrapper[4822]: I1124 14:37:38.173779 4822 scope.go:117] "RemoveContainer" containerID="b153e2957435bffbf591cf33e82e6ca36cb25843f38f81ff572770fd522f39e4" Nov 24 14:37:38 crc kubenswrapper[4822]: I1124 14:37:38.186897 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" podStartSLOduration=4.186881644 podStartE2EDuration="4.186881644s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:37:38.185145473 +0000 UTC m=+1095.301785950" watchObservedRunningTime="2025-11-24 14:37:38.186881644 +0000 UTC m=+1095.303522121" Nov 24 14:37:38 crc kubenswrapper[4822]: I1124 14:37:38.250170 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-hj58z"] Nov 24 14:37:38 crc kubenswrapper[4822]: I1124 14:37:38.282654 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-hj58z"] Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.188914 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a35b04-c993-40b5-8da8-8f19ce23a03b","Type":"ContainerStarted","Data":"45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b"} Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.189653 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a35b04-c993-40b5-8da8-8f19ce23a03b","Type":"ContainerStarted","Data":"f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749"} Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.189798 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerName="glance-log" containerID="cri-o://f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749" gracePeriod=30 Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.190367 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerName="glance-httpd" containerID="cri-o://45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b" gracePeriod=30 Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.223642 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"caf6e332-337b-4356-9bdf-d3ddf3d92e59","Type":"ContainerStarted","Data":"372f3becccacf884ecd253cbd86b5cd9d980e51e7ac9a249785de2c34260881c"} Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.230137 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.230119596 podStartE2EDuration="5.230119596s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:37:39.211597713 +0000 UTC m=+1096.328238190" watchObservedRunningTime="2025-11-24 14:37:39.230119596 +0000 UTC m=+1096.346760073" Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.722272 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9498898a-d1db-4c10-a06b-f2e0af5f8360" path="/var/lib/kubelet/pods/9498898a-d1db-4c10-a06b-f2e0af5f8360/volumes" Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.843261 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.990823 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-combined-ca-bundle\") pod \"00a35b04-c993-40b5-8da8-8f19ce23a03b\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.990908 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-public-tls-certs\") pod \"00a35b04-c993-40b5-8da8-8f19ce23a03b\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.990939 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-logs\") pod \"00a35b04-c993-40b5-8da8-8f19ce23a03b\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.991052 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"00a35b04-c993-40b5-8da8-8f19ce23a03b\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.991141 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-httpd-run\") pod \"00a35b04-c993-40b5-8da8-8f19ce23a03b\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.991227 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-config-data\") pod \"00a35b04-c993-40b5-8da8-8f19ce23a03b\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.991265 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-scripts\") pod \"00a35b04-c993-40b5-8da8-8f19ce23a03b\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " Nov 24 14:37:39 crc kubenswrapper[4822]: I1124 14:37:39.991287 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw6j8\" (UniqueName: \"kubernetes.io/projected/00a35b04-c993-40b5-8da8-8f19ce23a03b-kube-api-access-rw6j8\") pod \"00a35b04-c993-40b5-8da8-8f19ce23a03b\" (UID: \"00a35b04-c993-40b5-8da8-8f19ce23a03b\") " Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.006453 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "00a35b04-c993-40b5-8da8-8f19ce23a03b" (UID: "00a35b04-c993-40b5-8da8-8f19ce23a03b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.006986 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-logs" (OuterVolumeSpecName: "logs") pod "00a35b04-c993-40b5-8da8-8f19ce23a03b" (UID: "00a35b04-c993-40b5-8da8-8f19ce23a03b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.011492 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-scripts" (OuterVolumeSpecName: "scripts") pod "00a35b04-c993-40b5-8da8-8f19ce23a03b" (UID: "00a35b04-c993-40b5-8da8-8f19ce23a03b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.018409 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a35b04-c993-40b5-8da8-8f19ce23a03b-kube-api-access-rw6j8" (OuterVolumeSpecName: "kube-api-access-rw6j8") pod "00a35b04-c993-40b5-8da8-8f19ce23a03b" (UID: "00a35b04-c993-40b5-8da8-8f19ce23a03b"). InnerVolumeSpecName "kube-api-access-rw6j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.031356 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00a35b04-c993-40b5-8da8-8f19ce23a03b" (UID: "00a35b04-c993-40b5-8da8-8f19ce23a03b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.039879 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1" (OuterVolumeSpecName: "glance") pod "00a35b04-c993-40b5-8da8-8f19ce23a03b" (UID: "00a35b04-c993-40b5-8da8-8f19ce23a03b"). InnerVolumeSpecName "pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.063341 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "00a35b04-c993-40b5-8da8-8f19ce23a03b" (UID: "00a35b04-c993-40b5-8da8-8f19ce23a03b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.077983 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-config-data" (OuterVolumeSpecName: "config-data") pod "00a35b04-c993-40b5-8da8-8f19ce23a03b" (UID: "00a35b04-c993-40b5-8da8-8f19ce23a03b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.095877 4822 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.095909 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.095941 4822 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") on node \"crc\" " Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.095953 4822 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a35b04-c993-40b5-8da8-8f19ce23a03b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.095963 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.095970 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.095979 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw6j8\" (UniqueName: \"kubernetes.io/projected/00a35b04-c993-40b5-8da8-8f19ce23a03b-kube-api-access-rw6j8\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.095988 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a35b04-c993-40b5-8da8-8f19ce23a03b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.131614 4822 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.131731 4822 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1") on node "crc" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.197881 4822 reconciler_common.go:293] "Volume detached for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.241572 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"caf6e332-337b-4356-9bdf-d3ddf3d92e59","Type":"ContainerStarted","Data":"da04933bee0831fbbd177e8238f07a93128360c3da3bbadc60cebb7b244f2415"} Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.241705 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerName="glance-log" containerID="cri-o://372f3becccacf884ecd253cbd86b5cd9d980e51e7ac9a249785de2c34260881c" gracePeriod=30 Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.241733 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerName="glance-httpd" containerID="cri-o://da04933bee0831fbbd177e8238f07a93128360c3da3bbadc60cebb7b244f2415" gracePeriod=30 Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.244182 4822 generic.go:334] "Generic (PLEG): container finished" podID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerID="45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b" exitCode=143 Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.244223 4822 generic.go:334] "Generic (PLEG): container finished" podID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerID="f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749" exitCode=143 Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.244264 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a35b04-c993-40b5-8da8-8f19ce23a03b","Type":"ContainerDied","Data":"45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b"} Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.244293 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a35b04-c993-40b5-8da8-8f19ce23a03b","Type":"ContainerDied","Data":"f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749"} Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.244303 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a35b04-c993-40b5-8da8-8f19ce23a03b","Type":"ContainerDied","Data":"ec7e886af5719edd61ed865f9a3a1b69fac243c54e7fddf5e7e4b3c52516f3ee"} Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.244333 4822 scope.go:117] "RemoveContainer" containerID="45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.244443 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.253996 4822 generic.go:334] "Generic (PLEG): container finished" podID="1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" containerID="478c0129945058472624762055c6cbbcff7171196e7b8e7b05755afd0d7ab513" exitCode=0 Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.255146 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hbspf" event={"ID":"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f","Type":"ContainerDied","Data":"478c0129945058472624762055c6cbbcff7171196e7b8e7b05755afd0d7ab513"} Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.287457 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.287405248 podStartE2EDuration="6.287405248s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:37:40.267004721 +0000 UTC m=+1097.383645218" watchObservedRunningTime="2025-11-24 14:37:40.287405248 +0000 UTC m=+1097.404045725" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.294764 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.312488 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.325032 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:37:40 crc kubenswrapper[4822]: E1124 14:37:40.325476 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9498898a-d1db-4c10-a06b-f2e0af5f8360" containerName="init" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.325492 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9498898a-d1db-4c10-a06b-f2e0af5f8360" containerName="init" Nov 24 14:37:40 crc kubenswrapper[4822]: E1124 14:37:40.325506 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerName="glance-log" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.325513 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerName="glance-log" Nov 24 14:37:40 crc kubenswrapper[4822]: E1124 14:37:40.325536 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerName="glance-httpd" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.325542 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerName="glance-httpd" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.325728 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerName="glance-httpd" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.325746 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="9498898a-d1db-4c10-a06b-f2e0af5f8360" containerName="init" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.325759 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a35b04-c993-40b5-8da8-8f19ce23a03b" containerName="glance-log" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.326807 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.332670 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.332930 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.383376 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.506850 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25q7b\" (UniqueName: \"kubernetes.io/projected/ae3c17ab-662f-4507-a8c6-24a738cd3305-kube-api-access-25q7b\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.506915 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-config-data\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.506974 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.506995 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.507025 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.507055 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-scripts\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.507090 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.507192 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-logs\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.608684 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-logs\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.608743 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25q7b\" (UniqueName: \"kubernetes.io/projected/ae3c17ab-662f-4507-a8c6-24a738cd3305-kube-api-access-25q7b\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.608781 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-config-data\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.608865 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.608889 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.608942 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.608975 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-scripts\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.609035 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.609087 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-logs\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.611836 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.614496 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-config-data\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.615003 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.615031 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2ffde4dda1be13a8596eb6948dbe2a502fe570389c8cc35530979ae06145ae21/globalmount\"" pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.615770 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-scripts\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.617694 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.621504 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.630830 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25q7b\" (UniqueName: \"kubernetes.io/projected/ae3c17ab-662f-4507-a8c6-24a738cd3305-kube-api-access-25q7b\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.648439 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.667553 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:37:40 crc kubenswrapper[4822]: I1124 14:37:40.914440 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:37:40 crc kubenswrapper[4822]: E1124 14:37:40.914703 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:37:40 crc kubenswrapper[4822]: E1124 14:37:40.914763 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:37:40 crc kubenswrapper[4822]: E1124 14:37:40.914892 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:38:44.91485073 +0000 UTC m=+1162.031491207 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:37:41 crc kubenswrapper[4822]: I1124 14:37:41.265963 4822 generic.go:334] "Generic (PLEG): container finished" podID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerID="da04933bee0831fbbd177e8238f07a93128360c3da3bbadc60cebb7b244f2415" exitCode=0 Nov 24 14:37:41 crc kubenswrapper[4822]: I1124 14:37:41.265995 4822 generic.go:334] "Generic (PLEG): container finished" podID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerID="372f3becccacf884ecd253cbd86b5cd9d980e51e7ac9a249785de2c34260881c" exitCode=143 Nov 24 14:37:41 crc kubenswrapper[4822]: I1124 14:37:41.266005 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"caf6e332-337b-4356-9bdf-d3ddf3d92e59","Type":"ContainerDied","Data":"da04933bee0831fbbd177e8238f07a93128360c3da3bbadc60cebb7b244f2415"} Nov 24 14:37:41 crc kubenswrapper[4822]: I1124 14:37:41.266040 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"caf6e332-337b-4356-9bdf-d3ddf3d92e59","Type":"ContainerDied","Data":"372f3becccacf884ecd253cbd86b5cd9d980e51e7ac9a249785de2c34260881c"} Nov 24 14:37:41 crc kubenswrapper[4822]: I1124 14:37:41.578190 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:37:41 crc kubenswrapper[4822]: I1124 14:37:41.578617 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:37:41 crc kubenswrapper[4822]: I1124 14:37:41.717707 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a35b04-c993-40b5-8da8-8f19ce23a03b" path="/var/lib/kubelet/pods/00a35b04-c993-40b5-8da8-8f19ce23a03b/volumes" Nov 24 14:37:45 crc kubenswrapper[4822]: I1124 14:37:45.525737 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:37:45 crc kubenswrapper[4822]: I1124 14:37:45.592748 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-gkssz"] Nov 24 14:37:45 crc kubenswrapper[4822]: I1124 14:37:45.592983 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" podUID="981326ac-656e-410e-8016-7869af8ed9ea" containerName="dnsmasq-dns" containerID="cri-o://b77a3bdca919e625a1558a04b75e41b3591c9a2804450f405b7ab304139e6a2c" gracePeriod=10 Nov 24 14:37:46 crc kubenswrapper[4822]: I1124 14:37:46.339043 4822 generic.go:334] "Generic (PLEG): container finished" podID="981326ac-656e-410e-8016-7869af8ed9ea" containerID="b77a3bdca919e625a1558a04b75e41b3591c9a2804450f405b7ab304139e6a2c" exitCode=0 Nov 24 14:37:46 crc kubenswrapper[4822]: I1124 14:37:46.339079 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" event={"ID":"981326ac-656e-410e-8016-7869af8ed9ea","Type":"ContainerDied","Data":"b77a3bdca919e625a1558a04b75e41b3591c9a2804450f405b7ab304139e6a2c"} Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.350191 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hbspf" event={"ID":"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f","Type":"ContainerDied","Data":"81a4b0fcde9662886d8ecb66371e53becb64aef7b74978d0ebdaec2f72dde24f"} Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.350253 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81a4b0fcde9662886d8ecb66371e53becb64aef7b74978d0ebdaec2f72dde24f" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.434117 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.564989 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-config-data\") pod \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.565074 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nft6\" (UniqueName: \"kubernetes.io/projected/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-kube-api-access-9nft6\") pod \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.565258 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-scripts\") pod \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.565350 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-fernet-keys\") pod \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.565394 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-combined-ca-bundle\") pod \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.565468 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-credential-keys\") pod \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\" (UID: \"1726f74b-ca5e-43ac-9b0e-752ff7d63a9f\") " Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.570825 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" (UID: "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.572161 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-scripts" (OuterVolumeSpecName: "scripts") pod "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" (UID: "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.572229 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" (UID: "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.574502 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-kube-api-access-9nft6" (OuterVolumeSpecName: "kube-api-access-9nft6") pod "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" (UID: "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f"). InnerVolumeSpecName "kube-api-access-9nft6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.593168 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" (UID: "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.600279 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-config-data" (OuterVolumeSpecName: "config-data") pod "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" (UID: "1726f74b-ca5e-43ac-9b0e-752ff7d63a9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.667979 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.668033 4822 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.668053 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.668077 4822 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.668094 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:47 crc kubenswrapper[4822]: I1124 14:37:47.668111 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nft6\" (UniqueName: \"kubernetes.io/projected/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f-kube-api-access-9nft6\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.020522 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.178694 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-scripts\") pod \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.178786 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-logs\") pod \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.179434 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-logs" (OuterVolumeSpecName: "logs") pod "caf6e332-337b-4356-9bdf-d3ddf3d92e59" (UID: "caf6e332-337b-4356-9bdf-d3ddf3d92e59"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.179523 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-config-data\") pod \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.179893 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-internal-tls-certs\") pod \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.179961 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfwxc\" (UniqueName: \"kubernetes.io/projected/caf6e332-337b-4356-9bdf-d3ddf3d92e59-kube-api-access-hfwxc\") pod \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.179985 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-httpd-run\") pod \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.180085 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.180119 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-combined-ca-bundle\") pod \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\" (UID: \"caf6e332-337b-4356-9bdf-d3ddf3d92e59\") " Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.180724 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "caf6e332-337b-4356-9bdf-d3ddf3d92e59" (UID: "caf6e332-337b-4356-9bdf-d3ddf3d92e59"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.182549 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.182571 4822 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf6e332-337b-4356-9bdf-d3ddf3d92e59-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.187811 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-scripts" (OuterVolumeSpecName: "scripts") pod "caf6e332-337b-4356-9bdf-d3ddf3d92e59" (UID: "caf6e332-337b-4356-9bdf-d3ddf3d92e59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.188198 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf6e332-337b-4356-9bdf-d3ddf3d92e59-kube-api-access-hfwxc" (OuterVolumeSpecName: "kube-api-access-hfwxc") pod "caf6e332-337b-4356-9bdf-d3ddf3d92e59" (UID: "caf6e332-337b-4356-9bdf-d3ddf3d92e59"). InnerVolumeSpecName "kube-api-access-hfwxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.199059 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0" (OuterVolumeSpecName: "glance") pod "caf6e332-337b-4356-9bdf-d3ddf3d92e59" (UID: "caf6e332-337b-4356-9bdf-d3ddf3d92e59"). InnerVolumeSpecName "pvc-49360db5-9984-4b54-aac7-d578276794c0". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.212472 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "caf6e332-337b-4356-9bdf-d3ddf3d92e59" (UID: "caf6e332-337b-4356-9bdf-d3ddf3d92e59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.239420 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "caf6e332-337b-4356-9bdf-d3ddf3d92e59" (UID: "caf6e332-337b-4356-9bdf-d3ddf3d92e59"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.245960 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-config-data" (OuterVolumeSpecName: "config-data") pod "caf6e332-337b-4356-9bdf-d3ddf3d92e59" (UID: "caf6e332-337b-4356-9bdf-d3ddf3d92e59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.284648 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfwxc\" (UniqueName: \"kubernetes.io/projected/caf6e332-337b-4356-9bdf-d3ddf3d92e59-kube-api-access-hfwxc\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.284695 4822 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") on node \"crc\" " Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.284709 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.284720 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.284731 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.284741 4822 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/caf6e332-337b-4356-9bdf-d3ddf3d92e59-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.304761 4822 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.304888 4822 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-49360db5-9984-4b54-aac7-d578276794c0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0") on node "crc" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.360751 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hbspf" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.360949 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"caf6e332-337b-4356-9bdf-d3ddf3d92e59","Type":"ContainerDied","Data":"6a60118fcc880377c9a9dcc99f6e20baeefde3907de3742182f06770d9f406db"} Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.360977 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.388093 4822 reconciler_common.go:293] "Volume detached for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.411932 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.419250 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.456867 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:37:48 crc kubenswrapper[4822]: E1124 14:37:48.457625 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerName="glance-log" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.457663 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerName="glance-log" Nov 24 14:37:48 crc kubenswrapper[4822]: E1124 14:37:48.457702 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" containerName="keystone-bootstrap" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.457716 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" containerName="keystone-bootstrap" Nov 24 14:37:48 crc kubenswrapper[4822]: E1124 14:37:48.457762 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerName="glance-httpd" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.457776 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerName="glance-httpd" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.458135 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerName="glance-httpd" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.458180 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" containerName="glance-log" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.458248 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" containerName="keystone-bootstrap" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.460511 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.462454 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.463974 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.481916 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.543098 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hbspf"] Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.552496 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hbspf"] Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.593421 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.593494 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.593550 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw56s\" (UniqueName: \"kubernetes.io/projected/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-kube-api-access-fw56s\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.593662 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.593698 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.593755 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.593811 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.593852 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-logs\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.637913 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-7q6rz"] Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.639447 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.644237 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.645506 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.645746 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r2n9r" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.648854 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.662668 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7q6rz"] Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.696724 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.696809 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw56s\" (UniqueName: \"kubernetes.io/projected/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-kube-api-access-fw56s\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.696877 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.696902 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.696927 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.696967 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.696997 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-logs\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.697037 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.697277 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.699015 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-logs\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.700637 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.700672 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4ad6c7c6bc95d7d12c286d711c7d508fc5826003cdc7d089436fe350e33badd7/globalmount\"" pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.702960 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.703405 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.704278 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.712816 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.713415 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw56s\" (UniqueName: \"kubernetes.io/projected/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-kube-api-access-fw56s\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.744376 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.790777 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.798973 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-credential-keys\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.799019 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-config-data\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.799037 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-combined-ca-bundle\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.799078 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dswth\" (UniqueName: \"kubernetes.io/projected/858eb7c9-030e-49d8-b018-fdf41e979d20-kube-api-access-dswth\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.799169 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-scripts\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.799216 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-fernet-keys\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.900856 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-credential-keys\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.900902 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-config-data\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.900921 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-combined-ca-bundle\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.900950 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dswth\" (UniqueName: \"kubernetes.io/projected/858eb7c9-030e-49d8-b018-fdf41e979d20-kube-api-access-dswth\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.901020 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-scripts\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.901043 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-fernet-keys\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.909440 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-scripts\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.912034 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-config-data\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.915108 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-combined-ca-bundle\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.925289 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-fernet-keys\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.928911 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-credential-keys\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.950659 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dswth\" (UniqueName: \"kubernetes.io/projected/858eb7c9-030e-49d8-b018-fdf41e979d20-kube-api-access-dswth\") pod \"keystone-bootstrap-7q6rz\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:48 crc kubenswrapper[4822]: I1124 14:37:48.960568 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:37:49 crc kubenswrapper[4822]: I1124 14:37:49.716661 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1726f74b-ca5e-43ac-9b0e-752ff7d63a9f" path="/var/lib/kubelet/pods/1726f74b-ca5e-43ac-9b0e-752ff7d63a9f/volumes" Nov 24 14:37:49 crc kubenswrapper[4822]: I1124 14:37:49.717721 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caf6e332-337b-4356-9bdf-d3ddf3d92e59" path="/var/lib/kubelet/pods/caf6e332-337b-4356-9bdf-d3ddf3d92e59/volumes" Nov 24 14:37:50 crc kubenswrapper[4822]: I1124 14:37:50.291177 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" podUID="981326ac-656e-410e-8016-7869af8ed9ea" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: connect: connection refused" Nov 24 14:37:54 crc kubenswrapper[4822]: I1124 14:37:54.415017 4822 generic.go:334] "Generic (PLEG): container finished" podID="35fdf286-3217-480f-9522-3d146be7ceab" containerID="5ef5cc9c7238acea58a84c5a18958f784a38b03e541d20ea658682453ba33a37" exitCode=0 Nov 24 14:37:54 crc kubenswrapper[4822]: I1124 14:37:54.415099 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8rtpz" event={"ID":"35fdf286-3217-480f-9522-3d146be7ceab","Type":"ContainerDied","Data":"5ef5cc9c7238acea58a84c5a18958f784a38b03e541d20ea658682453ba33a37"} Nov 24 14:38:00 crc kubenswrapper[4822]: I1124 14:38:00.290321 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" podUID="981326ac-656e-410e-8016-7869af8ed9ea" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: i/o timeout" Nov 24 14:38:00 crc kubenswrapper[4822]: I1124 14:38:00.553490 4822 scope.go:117] "RemoveContainer" containerID="f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.186672 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.193451 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.388494 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-config\") pod \"35fdf286-3217-480f-9522-3d146be7ceab\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.388585 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-config\") pod \"981326ac-656e-410e-8016-7869af8ed9ea\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.388650 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flxjh\" (UniqueName: \"kubernetes.io/projected/981326ac-656e-410e-8016-7869af8ed9ea-kube-api-access-flxjh\") pod \"981326ac-656e-410e-8016-7869af8ed9ea\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.388696 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-dns-svc\") pod \"981326ac-656e-410e-8016-7869af8ed9ea\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.388811 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-nb\") pod \"981326ac-656e-410e-8016-7869af8ed9ea\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.388890 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjf24\" (UniqueName: \"kubernetes.io/projected/35fdf286-3217-480f-9522-3d146be7ceab-kube-api-access-fjf24\") pod \"35fdf286-3217-480f-9522-3d146be7ceab\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.388925 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-combined-ca-bundle\") pod \"35fdf286-3217-480f-9522-3d146be7ceab\" (UID: \"35fdf286-3217-480f-9522-3d146be7ceab\") " Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.388984 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-sb\") pod \"981326ac-656e-410e-8016-7869af8ed9ea\" (UID: \"981326ac-656e-410e-8016-7869af8ed9ea\") " Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.409094 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/981326ac-656e-410e-8016-7869af8ed9ea-kube-api-access-flxjh" (OuterVolumeSpecName: "kube-api-access-flxjh") pod "981326ac-656e-410e-8016-7869af8ed9ea" (UID: "981326ac-656e-410e-8016-7869af8ed9ea"). InnerVolumeSpecName "kube-api-access-flxjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.409871 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35fdf286-3217-480f-9522-3d146be7ceab-kube-api-access-fjf24" (OuterVolumeSpecName: "kube-api-access-fjf24") pod "35fdf286-3217-480f-9522-3d146be7ceab" (UID: "35fdf286-3217-480f-9522-3d146be7ceab"). InnerVolumeSpecName "kube-api-access-fjf24". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.421357 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35fdf286-3217-480f-9522-3d146be7ceab" (UID: "35fdf286-3217-480f-9522-3d146be7ceab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.431699 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-config" (OuterVolumeSpecName: "config") pod "35fdf286-3217-480f-9522-3d146be7ceab" (UID: "35fdf286-3217-480f-9522-3d146be7ceab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.456691 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-config" (OuterVolumeSpecName: "config") pod "981326ac-656e-410e-8016-7869af8ed9ea" (UID: "981326ac-656e-410e-8016-7869af8ed9ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.461785 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "981326ac-656e-410e-8016-7869af8ed9ea" (UID: "981326ac-656e-410e-8016-7869af8ed9ea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.464575 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "981326ac-656e-410e-8016-7869af8ed9ea" (UID: "981326ac-656e-410e-8016-7869af8ed9ea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.471856 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "981326ac-656e-410e-8016-7869af8ed9ea" (UID: "981326ac-656e-410e-8016-7869af8ed9ea"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.486062 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" event={"ID":"981326ac-656e-410e-8016-7869af8ed9ea","Type":"ContainerDied","Data":"f49ebb866eb886373277ffc7081dae0b33d8680b3f2378554f89ed3b471d7a7c"} Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.486325 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.487474 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8rtpz" event={"ID":"35fdf286-3217-480f-9522-3d146be7ceab","Type":"ContainerDied","Data":"dbc79fef58ca58767d123fd6668334e7824fecd708b4a8ea53ccf4f21559adab"} Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.487525 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbc79fef58ca58767d123fd6668334e7824fecd708b4a8ea53ccf4f21559adab" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.487552 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8rtpz" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.490651 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.490679 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjf24\" (UniqueName: \"kubernetes.io/projected/35fdf286-3217-480f-9522-3d146be7ceab-kube-api-access-fjf24\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.490692 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.490702 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.490712 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/35fdf286-3217-480f-9522-3d146be7ceab-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.490720 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.490728 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flxjh\" (UniqueName: \"kubernetes.io/projected/981326ac-656e-410e-8016-7869af8ed9ea-kube-api-access-flxjh\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.490738 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981326ac-656e-410e-8016-7869af8ed9ea-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.524786 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-gkssz"] Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.534996 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-gkssz"] Nov 24 14:38:01 crc kubenswrapper[4822]: I1124 14:38:01.717544 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="981326ac-656e-410e-8016-7869af8ed9ea" path="/var/lib/kubelet/pods/981326ac-656e-410e-8016-7869af8ed9ea/volumes" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.555930 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fb745b69-962h6"] Nov 24 14:38:02 crc kubenswrapper[4822]: E1124 14:38:02.556605 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981326ac-656e-410e-8016-7869af8ed9ea" containerName="init" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.556618 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="981326ac-656e-410e-8016-7869af8ed9ea" containerName="init" Nov 24 14:38:02 crc kubenswrapper[4822]: E1124 14:38:02.556631 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35fdf286-3217-480f-9522-3d146be7ceab" containerName="neutron-db-sync" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.556637 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="35fdf286-3217-480f-9522-3d146be7ceab" containerName="neutron-db-sync" Nov 24 14:38:02 crc kubenswrapper[4822]: E1124 14:38:02.556650 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981326ac-656e-410e-8016-7869af8ed9ea" containerName="dnsmasq-dns" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.556655 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="981326ac-656e-410e-8016-7869af8ed9ea" containerName="dnsmasq-dns" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.556829 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="981326ac-656e-410e-8016-7869af8ed9ea" containerName="dnsmasq-dns" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.556845 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="35fdf286-3217-480f-9522-3d146be7ceab" containerName="neutron-db-sync" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.566387 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.610723 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fb745b69-962h6"] Nov 24 14:38:02 crc kubenswrapper[4822]: E1124 14:38:02.689183 4822 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 24 14:38:02 crc kubenswrapper[4822]: E1124 14:38:02.689390 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r72z7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-fzrq8_openstack(7348b48b-1f84-439f-a049-bfc19935dbfc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:38:02 crc kubenswrapper[4822]: E1124 14:38:02.690824 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-fzrq8" podUID="7348b48b-1f84-439f-a049-bfc19935dbfc" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.709950 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d9f4b5fbd-wzx5t"] Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.711585 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.718727 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.718965 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.719073 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2mjb8" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.719174 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.720969 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-nb\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.721051 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-dns-svc\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.721309 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkqwt\" (UniqueName: \"kubernetes.io/projected/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-kube-api-access-nkqwt\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.721339 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-config\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.721362 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-sb\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.743600 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d9f4b5fbd-wzx5t"] Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.825219 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-ovndb-tls-certs\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.825461 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkqwt\" (UniqueName: \"kubernetes.io/projected/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-kube-api-access-nkqwt\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.825580 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-config\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.825652 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-config\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.825729 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-sb\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.825816 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-combined-ca-bundle\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.825931 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-nb\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.826011 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-httpd-config\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.826082 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-dns-svc\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.826169 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsnks\" (UniqueName: \"kubernetes.io/projected/e86a7914-dd9a-4b4e-abd7-86d055505255-kube-api-access-tsnks\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.827870 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-sb\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.828712 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-config\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.828727 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-nb\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.829415 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-dns-svc\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.846600 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkqwt\" (UniqueName: \"kubernetes.io/projected/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-kube-api-access-nkqwt\") pod \"dnsmasq-dns-fb745b69-962h6\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.902005 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.927642 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsnks\" (UniqueName: \"kubernetes.io/projected/e86a7914-dd9a-4b4e-abd7-86d055505255-kube-api-access-tsnks\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.927714 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-ovndb-tls-certs\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.927768 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-config\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.927811 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-combined-ca-bundle\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.928021 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-httpd-config\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.933127 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-config\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.933905 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-httpd-config\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.943927 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-combined-ca-bundle\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.947157 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsnks\" (UniqueName: \"kubernetes.io/projected/e86a7914-dd9a-4b4e-abd7-86d055505255-kube-api-access-tsnks\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:02 crc kubenswrapper[4822]: I1124 14:38:02.947162 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-ovndb-tls-certs\") pod \"neutron-5d9f4b5fbd-wzx5t\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:03 crc kubenswrapper[4822]: I1124 14:38:03.193810 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:03 crc kubenswrapper[4822]: E1124 14:38:03.555130 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-fzrq8" podUID="7348b48b-1f84-439f-a049-bfc19935dbfc" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.770183 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cdd657c7c-xmwlh"] Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.777289 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.782018 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cdd657c7c-xmwlh"] Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.786998 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.787224 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.868540 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-internal-tls-certs\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.868611 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-httpd-config\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.868790 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-combined-ca-bundle\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.868841 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-ovndb-tls-certs\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.869083 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-config\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.869257 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-public-tls-certs\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.869324 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk8n5\" (UniqueName: \"kubernetes.io/projected/ef13e0bb-79cf-492c-880b-22a79fc0045b-kube-api-access-kk8n5\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.970660 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-config\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.970744 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-public-tls-certs\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.970780 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk8n5\" (UniqueName: \"kubernetes.io/projected/ef13e0bb-79cf-492c-880b-22a79fc0045b-kube-api-access-kk8n5\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.970826 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-internal-tls-certs\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.970849 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-httpd-config\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.970881 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-combined-ca-bundle\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.970899 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-ovndb-tls-certs\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.979937 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-ovndb-tls-certs\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.980263 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-internal-tls-certs\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.981619 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-config\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.981651 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-combined-ca-bundle\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.987611 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-public-tls-certs\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:04 crc kubenswrapper[4822]: I1124 14:38:04.999085 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ef13e0bb-79cf-492c-880b-22a79fc0045b-httpd-config\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.002958 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk8n5\" (UniqueName: \"kubernetes.io/projected/ef13e0bb-79cf-492c-880b-22a79fc0045b-kube-api-access-kk8n5\") pod \"neutron-6cdd657c7c-xmwlh\" (UID: \"ef13e0bb-79cf-492c-880b-22a79fc0045b\") " pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.098899 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.291133 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-gkssz" podUID="981326ac-656e-410e-8016-7869af8ed9ea" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: i/o timeout" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.807456 4822 scope.go:117] "RemoveContainer" containerID="45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b" Nov 24 14:38:05 crc kubenswrapper[4822]: E1124 14:38:05.808385 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b\": container with ID starting with 45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b not found: ID does not exist" containerID="45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.808542 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b"} err="failed to get container status \"45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b\": rpc error: code = NotFound desc = could not find container \"45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b\": container with ID starting with 45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b not found: ID does not exist" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.808580 4822 scope.go:117] "RemoveContainer" containerID="f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749" Nov 24 14:38:05 crc kubenswrapper[4822]: E1124 14:38:05.809116 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749\": container with ID starting with f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749 not found: ID does not exist" containerID="f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.809162 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749"} err="failed to get container status \"f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749\": rpc error: code = NotFound desc = could not find container \"f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749\": container with ID starting with f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749 not found: ID does not exist" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.809191 4822 scope.go:117] "RemoveContainer" containerID="45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.809864 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b"} err="failed to get container status \"45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b\": rpc error: code = NotFound desc = could not find container \"45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b\": container with ID starting with 45fdb526d4748d8037468777f4d7d10e3c9cff7da1afad57cf1b90ab0050065b not found: ID does not exist" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.809970 4822 scope.go:117] "RemoveContainer" containerID="f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.810508 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749"} err="failed to get container status \"f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749\": rpc error: code = NotFound desc = could not find container \"f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749\": container with ID starting with f3dd5cbdd80c2c9bc5c5fdc14029e8e2fabf975e1205000f3a7b600ac58f8749 not found: ID does not exist" Nov 24 14:38:05 crc kubenswrapper[4822]: I1124 14:38:05.810531 4822 scope.go:117] "RemoveContainer" containerID="da04933bee0831fbbd177e8238f07a93128360c3da3bbadc60cebb7b244f2415" Nov 24 14:38:06 crc kubenswrapper[4822]: I1124 14:38:06.302442 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:38:08 crc kubenswrapper[4822]: I1124 14:38:08.665365 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7q6rz"] Nov 24 14:38:08 crc kubenswrapper[4822]: I1124 14:38:08.865067 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:38:08 crc kubenswrapper[4822]: W1124 14:38:08.964615 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae3c17ab_662f_4507_a8c6_24a738cd3305.slice/crio-02af026aa88c2847acb335ad83d45fddebf5b1429c38def0ca42fc61eae931e7 WatchSource:0}: Error finding container 02af026aa88c2847acb335ad83d45fddebf5b1429c38def0ca42fc61eae931e7: Status 404 returned error can't find the container with id 02af026aa88c2847acb335ad83d45fddebf5b1429c38def0ca42fc61eae931e7 Nov 24 14:38:08 crc kubenswrapper[4822]: W1124 14:38:08.978226 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod858eb7c9_030e_49d8_b018_fdf41e979d20.slice/crio-d90c52fb74c780df1364b75fe6258a3c392d5125abd073d943f6b2f0557c1379 WatchSource:0}: Error finding container d90c52fb74c780df1364b75fe6258a3c392d5125abd073d943f6b2f0557c1379: Status 404 returned error can't find the container with id d90c52fb74c780df1364b75fe6258a3c392d5125abd073d943f6b2f0557c1379 Nov 24 14:38:08 crc kubenswrapper[4822]: W1124 14:38:08.979368 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb16d0ea_2a9a_4b31_897c_eb6938531d2e.slice/crio-5c501dccd586a4ed774543deb55537bc98e6b79be183d21e4a32fb96690a62d0 WatchSource:0}: Error finding container 5c501dccd586a4ed774543deb55537bc98e6b79be183d21e4a32fb96690a62d0: Status 404 returned error can't find the container with id 5c501dccd586a4ed774543deb55537bc98e6b79be183d21e4a32fb96690a62d0 Nov 24 14:38:08 crc kubenswrapper[4822]: I1124 14:38:08.987513 4822 scope.go:117] "RemoveContainer" containerID="372f3becccacf884ecd253cbd86b5cd9d980e51e7ac9a249785de2c34260881c" Nov 24 14:38:09 crc kubenswrapper[4822]: E1124 14:38:09.011753 4822 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Nov 24 14:38:09 crc kubenswrapper[4822]: E1124 14:38:09.012335 4822 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Nov 24 14:38:09 crc kubenswrapper[4822]: E1124 14:38:09.012470 4822 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cloudkitty-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CloudKittyPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:CloudKittyPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:cloudkitty-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:certs,ReadOnly:true,MountPath:/var/lib/openstack/loki-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rx97z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42406,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-db-sync-s6qvf_openstack(f0ad3a91-d60c-40a0-a285-c5a9615fffe0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:38:09 crc kubenswrapper[4822]: E1124 14:38:09.013680 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cloudkitty-db-sync-s6qvf" podUID="f0ad3a91-d60c-40a0-a285-c5a9615fffe0" Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.324042 4822 scope.go:117] "RemoveContainer" containerID="b77a3bdca919e625a1558a04b75e41b3591c9a2804450f405b7ab304139e6a2c" Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.371607 4822 scope.go:117] "RemoveContainer" containerID="99bc53f4c11169da9d920d934dbbce72d43150e8cc269751b59db4fa9c0e621d" Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.476432 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fb745b69-962h6"] Nov 24 14:38:09 crc kubenswrapper[4822]: W1124 14:38:09.505248 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f0b9893_50bf_4550_a893_bd82d7e4ecc3.slice/crio-bfbd4d0e928c869f7dd1d3c71fb58a75ce9775d395299d17582d338e34cef78a WatchSource:0}: Error finding container bfbd4d0e928c869f7dd1d3c71fb58a75ce9775d395299d17582d338e34cef78a: Status 404 returned error can't find the container with id bfbd4d0e928c869f7dd1d3c71fb58a75ce9775d395299d17582d338e34cef78a Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.563436 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d9f4b5fbd-wzx5t"] Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.640066 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb16d0ea-2a9a-4b31-897c-eb6938531d2e","Type":"ContainerStarted","Data":"5c501dccd586a4ed774543deb55537bc98e6b79be183d21e4a32fb96690a62d0"} Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.647593 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9f4b5fbd-wzx5t" event={"ID":"e86a7914-dd9a-4b4e-abd7-86d055505255","Type":"ContainerStarted","Data":"f76b7c7767ce3d8d3322fa4494dc9ed25211873c367261fae919193a8454d8cf"} Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.652226 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7q6rz" event={"ID":"858eb7c9-030e-49d8-b018-fdf41e979d20","Type":"ContainerStarted","Data":"06c06097721e5f3337bcd2702a2f65c9e5e376dfef37c33d0b09c83d8085e70a"} Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.652268 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7q6rz" event={"ID":"858eb7c9-030e-49d8-b018-fdf41e979d20","Type":"ContainerStarted","Data":"d90c52fb74c780df1364b75fe6258a3c392d5125abd073d943f6b2f0557c1379"} Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.663023 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fb745b69-962h6" event={"ID":"3f0b9893-50bf-4550-a893-bd82d7e4ecc3","Type":"ContainerStarted","Data":"bfbd4d0e928c869f7dd1d3c71fb58a75ce9775d395299d17582d338e34cef78a"} Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.674403 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cdd657c7c-xmwlh"] Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.674573 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-7q6rz" podStartSLOduration=21.674556732 podStartE2EDuration="21.674556732s" podCreationTimestamp="2025-11-24 14:37:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:09.669051022 +0000 UTC m=+1126.785691499" watchObservedRunningTime="2025-11-24 14:38:09.674556732 +0000 UTC m=+1126.791197209" Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.682897 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ae3c17ab-662f-4507-a8c6-24a738cd3305","Type":"ContainerStarted","Data":"02af026aa88c2847acb335ad83d45fddebf5b1429c38def0ca42fc61eae931e7"} Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.722944 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-fxf96" podStartSLOduration=9.018172378 podStartE2EDuration="35.722929526s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="2025-11-24 14:37:35.99214546 +0000 UTC m=+1093.108785937" lastFinishedPulling="2025-11-24 14:38:02.696902608 +0000 UTC m=+1119.813543085" observedRunningTime="2025-11-24 14:38:09.715643431 +0000 UTC m=+1126.832283908" watchObservedRunningTime="2025-11-24 14:38:09.722929526 +0000 UTC m=+1126.839570003" Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.729569 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fxf96" event={"ID":"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd","Type":"ContainerStarted","Data":"0fc918670c183ca24392e2580eca485a0ebecfebdd000e8956d4ff66d7f2bacb"} Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.732900 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-227wq" event={"ID":"42faa16a-d5df-4002-94bc-dd9b727ad202","Type":"ContainerStarted","Data":"0aac29ea10e6cfe52c197751f5798769a1e2f22a4e15a51e51eba8666a586266"} Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.737393 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerStarted","Data":"f9a32f0cc45408679a748bfb9c5fe147c7e9c92bf6130f47e7e57fe7e0b2eaa7"} Nov 24 14:38:09 crc kubenswrapper[4822]: E1124 14:38:09.738701 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current\\\"\"" pod="openstack/cloudkitty-db-sync-s6qvf" podUID="f0ad3a91-d60c-40a0-a285-c5a9615fffe0" Nov 24 14:38:09 crc kubenswrapper[4822]: I1124 14:38:09.753497 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-227wq" podStartSLOduration=11.42053377 podStartE2EDuration="35.753478059s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="2025-11-24 14:37:36.221981813 +0000 UTC m=+1093.338622290" lastFinishedPulling="2025-11-24 14:38:00.554926092 +0000 UTC m=+1117.671566579" observedRunningTime="2025-11-24 14:38:09.747530555 +0000 UTC m=+1126.864171032" watchObservedRunningTime="2025-11-24 14:38:09.753478059 +0000 UTC m=+1126.870118536" Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.773023 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdd657c7c-xmwlh" event={"ID":"ef13e0bb-79cf-492c-880b-22a79fc0045b","Type":"ContainerStarted","Data":"12289baa7e017698b0030a416d77850600b875c9adcd697856f6ca2288ef098c"} Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.773306 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdd657c7c-xmwlh" event={"ID":"ef13e0bb-79cf-492c-880b-22a79fc0045b","Type":"ContainerStarted","Data":"d8a45fe27321ca352eaa1c6131d757aa81c5f73d55fa767507feba6486fc2396"} Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.773316 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdd657c7c-xmwlh" event={"ID":"ef13e0bb-79cf-492c-880b-22a79fc0045b","Type":"ContainerStarted","Data":"aa8382d87cf1cefc966a82c4d979b955e30151b82139ef73d3584410732ad5ca"} Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.774644 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.776844 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb16d0ea-2a9a-4b31-897c-eb6938531d2e","Type":"ContainerStarted","Data":"cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e"} Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.778773 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ae3c17ab-662f-4507-a8c6-24a738cd3305","Type":"ContainerStarted","Data":"09bdf186b9cc111d712361299a44adf688cfdffe95daf750743a3bd73aa6ebc8"} Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.780710 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9f4b5fbd-wzx5t" event={"ID":"e86a7914-dd9a-4b4e-abd7-86d055505255","Type":"ContainerStarted","Data":"786e0d43cffd1b10ee7645509434625f6b501a6e88bc6be1b2f820f28ee2dcd8"} Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.780753 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9f4b5fbd-wzx5t" event={"ID":"e86a7914-dd9a-4b4e-abd7-86d055505255","Type":"ContainerStarted","Data":"f5de7618b4c29ab4b842e9d844eae59251eceff2206035110af1d83066023756"} Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.780841 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.784603 4822 generic.go:334] "Generic (PLEG): container finished" podID="3f0b9893-50bf-4550-a893-bd82d7e4ecc3" containerID="d9f4f5ac8fdfc931f03470a4d5f62861c24ce40b5b9df4e438cf6a21a4b1393c" exitCode=0 Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.785719 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fb745b69-962h6" event={"ID":"3f0b9893-50bf-4550-a893-bd82d7e4ecc3","Type":"ContainerDied","Data":"d9f4f5ac8fdfc931f03470a4d5f62861c24ce40b5b9df4e438cf6a21a4b1393c"} Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.791445 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cdd657c7c-xmwlh" podStartSLOduration=6.791435147 podStartE2EDuration="6.791435147s" podCreationTimestamp="2025-11-24 14:38:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:10.788986951 +0000 UTC m=+1127.905627428" watchObservedRunningTime="2025-11-24 14:38:10.791435147 +0000 UTC m=+1127.908075624" Nov 24 14:38:10 crc kubenswrapper[4822]: I1124 14:38:10.842859 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d9f4b5fbd-wzx5t" podStartSLOduration=8.842843804 podStartE2EDuration="8.842843804s" podCreationTimestamp="2025-11-24 14:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:10.841921856 +0000 UTC m=+1127.958562333" watchObservedRunningTime="2025-11-24 14:38:10.842843804 +0000 UTC m=+1127.959484281" Nov 24 14:38:11 crc kubenswrapper[4822]: I1124 14:38:11.577389 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:38:11 crc kubenswrapper[4822]: I1124 14:38:11.577716 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:38:11 crc kubenswrapper[4822]: I1124 14:38:11.793901 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb16d0ea-2a9a-4b31-897c-eb6938531d2e","Type":"ContainerStarted","Data":"6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06"} Nov 24 14:38:11 crc kubenswrapper[4822]: I1124 14:38:11.795650 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ae3c17ab-662f-4507-a8c6-24a738cd3305","Type":"ContainerStarted","Data":"8fba0d1620c0c0b27290a0c8006fbcd806f65acec359a6e086370786aab000b1"} Nov 24 14:38:11 crc kubenswrapper[4822]: I1124 14:38:11.797049 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fb745b69-962h6" event={"ID":"3f0b9893-50bf-4550-a893-bd82d7e4ecc3","Type":"ContainerStarted","Data":"a6d0ce151b87bdef67f37e6d46b107fd8db4222d090cef61bea25935634d62ab"} Nov 24 14:38:11 crc kubenswrapper[4822]: I1124 14:38:11.821648 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=23.821624135 podStartE2EDuration="23.821624135s" podCreationTimestamp="2025-11-24 14:37:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:11.816649152 +0000 UTC m=+1128.933289639" watchObservedRunningTime="2025-11-24 14:38:11.821624135 +0000 UTC m=+1128.938264612" Nov 24 14:38:11 crc kubenswrapper[4822]: I1124 14:38:11.849566 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=31.849519427 podStartE2EDuration="31.849519427s" podCreationTimestamp="2025-11-24 14:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:11.845702669 +0000 UTC m=+1128.962343146" watchObservedRunningTime="2025-11-24 14:38:11.849519427 +0000 UTC m=+1128.966159904" Nov 24 14:38:11 crc kubenswrapper[4822]: I1124 14:38:11.882176 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fb745b69-962h6" podStartSLOduration=9.882161725 podStartE2EDuration="9.882161725s" podCreationTimestamp="2025-11-24 14:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:11.879609146 +0000 UTC m=+1128.996249643" watchObservedRunningTime="2025-11-24 14:38:11.882161725 +0000 UTC m=+1128.998802202" Nov 24 14:38:12 crc kubenswrapper[4822]: I1124 14:38:12.808283 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:14 crc kubenswrapper[4822]: I1124 14:38:14.836140 4822 generic.go:334] "Generic (PLEG): container finished" podID="858eb7c9-030e-49d8-b018-fdf41e979d20" containerID="06c06097721e5f3337bcd2702a2f65c9e5e376dfef37c33d0b09c83d8085e70a" exitCode=0 Nov 24 14:38:14 crc kubenswrapper[4822]: I1124 14:38:14.836230 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7q6rz" event={"ID":"858eb7c9-030e-49d8-b018-fdf41e979d20","Type":"ContainerDied","Data":"06c06097721e5f3337bcd2702a2f65c9e5e376dfef37c33d0b09c83d8085e70a"} Nov 24 14:38:14 crc kubenswrapper[4822]: I1124 14:38:14.839192 4822 generic.go:334] "Generic (PLEG): container finished" podID="42faa16a-d5df-4002-94bc-dd9b727ad202" containerID="0aac29ea10e6cfe52c197751f5798769a1e2f22a4e15a51e51eba8666a586266" exitCode=0 Nov 24 14:38:14 crc kubenswrapper[4822]: I1124 14:38:14.839227 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-227wq" event={"ID":"42faa16a-d5df-4002-94bc-dd9b727ad202","Type":"ContainerDied","Data":"0aac29ea10e6cfe52c197751f5798769a1e2f22a4e15a51e51eba8666a586266"} Nov 24 14:38:14 crc kubenswrapper[4822]: I1124 14:38:14.841611 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerStarted","Data":"13e53dd3c7375c0c69232d41aa2e75146658d67590f39698fddd3dc82d27f7a8"} Nov 24 14:38:14 crc kubenswrapper[4822]: I1124 14:38:14.843158 4822 generic.go:334] "Generic (PLEG): container finished" podID="5cebbef1-88ea-4dd3-9a1d-3900d2d160bd" containerID="0fc918670c183ca24392e2580eca485a0ebecfebdd000e8956d4ff66d7f2bacb" exitCode=0 Nov 24 14:38:14 crc kubenswrapper[4822]: I1124 14:38:14.843199 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fxf96" event={"ID":"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd","Type":"ContainerDied","Data":"0fc918670c183ca24392e2580eca485a0ebecfebdd000e8956d4ff66d7f2bacb"} Nov 24 14:38:17 crc kubenswrapper[4822]: I1124 14:38:17.904329 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.014377 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gmz62"] Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.014704 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" podUID="d2cfadc6-5078-4554-849b-c013b8fa0a2e" containerName="dnsmasq-dns" containerID="cri-o://4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae" gracePeriod=10 Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.635270 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.639572 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fxf96" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.645927 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-227wq" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.730749 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42faa16a-d5df-4002-94bc-dd9b727ad202-logs\") pod \"42faa16a-d5df-4002-94bc-dd9b727ad202\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.730784 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-credential-keys\") pod \"858eb7c9-030e-49d8-b018-fdf41e979d20\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.730804 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnz6m\" (UniqueName: \"kubernetes.io/projected/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-kube-api-access-gnz6m\") pod \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.730826 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-combined-ca-bundle\") pod \"858eb7c9-030e-49d8-b018-fdf41e979d20\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.730995 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dswth\" (UniqueName: \"kubernetes.io/projected/858eb7c9-030e-49d8-b018-fdf41e979d20-kube-api-access-dswth\") pod \"858eb7c9-030e-49d8-b018-fdf41e979d20\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731033 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz25l\" (UniqueName: \"kubernetes.io/projected/42faa16a-d5df-4002-94bc-dd9b727ad202-kube-api-access-dz25l\") pod \"42faa16a-d5df-4002-94bc-dd9b727ad202\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731054 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-scripts\") pod \"42faa16a-d5df-4002-94bc-dd9b727ad202\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731083 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-db-sync-config-data\") pod \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731099 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-config-data\") pod \"858eb7c9-030e-49d8-b018-fdf41e979d20\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731149 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-combined-ca-bundle\") pod \"42faa16a-d5df-4002-94bc-dd9b727ad202\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731284 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-combined-ca-bundle\") pod \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\" (UID: \"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731318 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-config-data\") pod \"42faa16a-d5df-4002-94bc-dd9b727ad202\" (UID: \"42faa16a-d5df-4002-94bc-dd9b727ad202\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731335 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-scripts\") pod \"858eb7c9-030e-49d8-b018-fdf41e979d20\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731354 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-fernet-keys\") pod \"858eb7c9-030e-49d8-b018-fdf41e979d20\" (UID: \"858eb7c9-030e-49d8-b018-fdf41e979d20\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.731849 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42faa16a-d5df-4002-94bc-dd9b727ad202-logs" (OuterVolumeSpecName: "logs") pod "42faa16a-d5df-4002-94bc-dd9b727ad202" (UID: "42faa16a-d5df-4002-94bc-dd9b727ad202"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.735787 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42faa16a-d5df-4002-94bc-dd9b727ad202-kube-api-access-dz25l" (OuterVolumeSpecName: "kube-api-access-dz25l") pod "42faa16a-d5df-4002-94bc-dd9b727ad202" (UID: "42faa16a-d5df-4002-94bc-dd9b727ad202"). InnerVolumeSpecName "kube-api-access-dz25l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.742981 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.744350 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "858eb7c9-030e-49d8-b018-fdf41e979d20" (UID: "858eb7c9-030e-49d8-b018-fdf41e979d20"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.744683 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-scripts" (OuterVolumeSpecName: "scripts") pod "42faa16a-d5df-4002-94bc-dd9b727ad202" (UID: "42faa16a-d5df-4002-94bc-dd9b727ad202"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.744760 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-kube-api-access-gnz6m" (OuterVolumeSpecName: "kube-api-access-gnz6m") pod "5cebbef1-88ea-4dd3-9a1d-3900d2d160bd" (UID: "5cebbef1-88ea-4dd3-9a1d-3900d2d160bd"). InnerVolumeSpecName "kube-api-access-gnz6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.745420 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "858eb7c9-030e-49d8-b018-fdf41e979d20" (UID: "858eb7c9-030e-49d8-b018-fdf41e979d20"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.745770 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/858eb7c9-030e-49d8-b018-fdf41e979d20-kube-api-access-dswth" (OuterVolumeSpecName: "kube-api-access-dswth") pod "858eb7c9-030e-49d8-b018-fdf41e979d20" (UID: "858eb7c9-030e-49d8-b018-fdf41e979d20"). InnerVolumeSpecName "kube-api-access-dswth". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.748015 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-scripts" (OuterVolumeSpecName: "scripts") pod "858eb7c9-030e-49d8-b018-fdf41e979d20" (UID: "858eb7c9-030e-49d8-b018-fdf41e979d20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.764717 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5cebbef1-88ea-4dd3-9a1d-3900d2d160bd" (UID: "5cebbef1-88ea-4dd3-9a1d-3900d2d160bd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.779049 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-config-data" (OuterVolumeSpecName: "config-data") pod "858eb7c9-030e-49d8-b018-fdf41e979d20" (UID: "858eb7c9-030e-49d8-b018-fdf41e979d20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.795708 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.795789 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.795803 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.795812 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.807528 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "858eb7c9-030e-49d8-b018-fdf41e979d20" (UID: "858eb7c9-030e-49d8-b018-fdf41e979d20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.821880 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-config-data" (OuterVolumeSpecName: "config-data") pod "42faa16a-d5df-4002-94bc-dd9b727ad202" (UID: "42faa16a-d5df-4002-94bc-dd9b727ad202"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.832476 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-nb\") pod \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.832605 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-dns-svc\") pod \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.832646 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-sb\") pod \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.832813 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k9qd\" (UniqueName: \"kubernetes.io/projected/d2cfadc6-5078-4554-849b-c013b8fa0a2e-kube-api-access-6k9qd\") pod \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.832873 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-config\") pod \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\" (UID: \"d2cfadc6-5078-4554-849b-c013b8fa0a2e\") " Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833321 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833336 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833345 4822 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833355 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42faa16a-d5df-4002-94bc-dd9b727ad202-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833363 4822 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833372 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnz6m\" (UniqueName: \"kubernetes.io/projected/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-kube-api-access-gnz6m\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833380 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833389 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dswth\" (UniqueName: \"kubernetes.io/projected/858eb7c9-030e-49d8-b018-fdf41e979d20-kube-api-access-dswth\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833397 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz25l\" (UniqueName: \"kubernetes.io/projected/42faa16a-d5df-4002-94bc-dd9b727ad202-kube-api-access-dz25l\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833406 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833414 4822 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.833423 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/858eb7c9-030e-49d8-b018-fdf41e979d20-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.846131 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2cfadc6-5078-4554-849b-c013b8fa0a2e-kube-api-access-6k9qd" (OuterVolumeSpecName: "kube-api-access-6k9qd") pod "d2cfadc6-5078-4554-849b-c013b8fa0a2e" (UID: "d2cfadc6-5078-4554-849b-c013b8fa0a2e"). InnerVolumeSpecName "kube-api-access-6k9qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.846231 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.846524 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.861696 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cebbef1-88ea-4dd3-9a1d-3900d2d160bd" (UID: "5cebbef1-88ea-4dd3-9a1d-3900d2d160bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.871439 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42faa16a-d5df-4002-94bc-dd9b727ad202" (UID: "42faa16a-d5df-4002-94bc-dd9b727ad202"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.891825 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fxf96" event={"ID":"5cebbef1-88ea-4dd3-9a1d-3900d2d160bd","Type":"ContainerDied","Data":"9d5b49489757cce5841d9b4c6364c2f74eada35a8068544cca3572dd309628e3"} Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.891840 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fxf96" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.891863 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d5b49489757cce5841d9b4c6364c2f74eada35a8068544cca3572dd309628e3" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.899817 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7q6rz" event={"ID":"858eb7c9-030e-49d8-b018-fdf41e979d20","Type":"ContainerDied","Data":"d90c52fb74c780df1364b75fe6258a3c392d5125abd073d943f6b2f0557c1379"} Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.899855 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d90c52fb74c780df1364b75fe6258a3c392d5125abd073d943f6b2f0557c1379" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.899905 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7q6rz" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.912365 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-227wq" event={"ID":"42faa16a-d5df-4002-94bc-dd9b727ad202","Type":"ContainerDied","Data":"e42521fcb2ef0acae9391961e55ec0f9aef8f204a0043b83ab411828b5c088ee"} Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.912406 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e42521fcb2ef0acae9391961e55ec0f9aef8f204a0043b83ab411828b5c088ee" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.912474 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-227wq" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.918502 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-config" (OuterVolumeSpecName: "config") pod "d2cfadc6-5078-4554-849b-c013b8fa0a2e" (UID: "d2cfadc6-5078-4554-849b-c013b8fa0a2e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.922311 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d2cfadc6-5078-4554-849b-c013b8fa0a2e" (UID: "d2cfadc6-5078-4554-849b-c013b8fa0a2e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.927925 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerStarted","Data":"6fd3b5a8ca9cfaa068d4131029730208157f302f41fbcd2df94c39285f3a9420"} Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.928288 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d2cfadc6-5078-4554-849b-c013b8fa0a2e" (UID: "d2cfadc6-5078-4554-849b-c013b8fa0a2e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.932617 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d2cfadc6-5078-4554-849b-c013b8fa0a2e" (UID: "d2cfadc6-5078-4554-849b-c013b8fa0a2e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.937625 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42faa16a-d5df-4002-94bc-dd9b727ad202-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.937648 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k9qd\" (UniqueName: \"kubernetes.io/projected/d2cfadc6-5078-4554-849b-c013b8fa0a2e-kube-api-access-6k9qd\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.937658 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.937669 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.937679 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.937687 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.937696 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2cfadc6-5078-4554-849b-c013b8fa0a2e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.942239 4822 generic.go:334] "Generic (PLEG): container finished" podID="d2cfadc6-5078-4554-849b-c013b8fa0a2e" containerID="4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae" exitCode=0 Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.943380 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.943406 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" event={"ID":"d2cfadc6-5078-4554-849b-c013b8fa0a2e","Type":"ContainerDied","Data":"4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae"} Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.943466 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-gmz62" event={"ID":"d2cfadc6-5078-4554-849b-c013b8fa0a2e","Type":"ContainerDied","Data":"667266956e195cff7e79d05e9e920e4591b7f186a798de69a37eeda0c649a2fc"} Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.943485 4822 scope.go:117] "RemoveContainer" containerID="4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.979937 4822 scope.go:117] "RemoveContainer" containerID="4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9" Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.981382 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gmz62"] Nov 24 14:38:18 crc kubenswrapper[4822]: I1124 14:38:18.991630 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gmz62"] Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.008653 4822 scope.go:117] "RemoveContainer" containerID="4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae" Nov 24 14:38:19 crc kubenswrapper[4822]: E1124 14:38:19.009122 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae\": container with ID starting with 4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae not found: ID does not exist" containerID="4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.009163 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae"} err="failed to get container status \"4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae\": rpc error: code = NotFound desc = could not find container \"4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae\": container with ID starting with 4326e1e35a1fb8dd131dfcbabfcb4a09b1c7dd02a4485396b1aed5690f1615ae not found: ID does not exist" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.009223 4822 scope.go:117] "RemoveContainer" containerID="4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9" Nov 24 14:38:19 crc kubenswrapper[4822]: E1124 14:38:19.009839 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9\": container with ID starting with 4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9 not found: ID does not exist" containerID="4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.009862 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9"} err="failed to get container status \"4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9\": rpc error: code = NotFound desc = could not find container \"4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9\": container with ID starting with 4470d1abd92efb96e236714b2bafb808bfbb12fd70bbf38ab79e4c0f6b78c8d9 not found: ID does not exist" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.716780 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2cfadc6-5078-4554-849b-c013b8fa0a2e" path="/var/lib/kubelet/pods/d2cfadc6-5078-4554-849b-c013b8fa0a2e/volumes" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.759639 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7f958c4db4-6hhpg"] Nov 24 14:38:19 crc kubenswrapper[4822]: E1124 14:38:19.760087 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42faa16a-d5df-4002-94bc-dd9b727ad202" containerName="placement-db-sync" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.760109 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="42faa16a-d5df-4002-94bc-dd9b727ad202" containerName="placement-db-sync" Nov 24 14:38:19 crc kubenswrapper[4822]: E1124 14:38:19.760130 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2cfadc6-5078-4554-849b-c013b8fa0a2e" containerName="dnsmasq-dns" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.760139 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2cfadc6-5078-4554-849b-c013b8fa0a2e" containerName="dnsmasq-dns" Nov 24 14:38:19 crc kubenswrapper[4822]: E1124 14:38:19.760153 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cebbef1-88ea-4dd3-9a1d-3900d2d160bd" containerName="barbican-db-sync" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.760160 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cebbef1-88ea-4dd3-9a1d-3900d2d160bd" containerName="barbican-db-sync" Nov 24 14:38:19 crc kubenswrapper[4822]: E1124 14:38:19.760179 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2cfadc6-5078-4554-849b-c013b8fa0a2e" containerName="init" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.760186 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2cfadc6-5078-4554-849b-c013b8fa0a2e" containerName="init" Nov 24 14:38:19 crc kubenswrapper[4822]: E1124 14:38:19.760224 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858eb7c9-030e-49d8-b018-fdf41e979d20" containerName="keystone-bootstrap" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.760232 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="858eb7c9-030e-49d8-b018-fdf41e979d20" containerName="keystone-bootstrap" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.760450 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cebbef1-88ea-4dd3-9a1d-3900d2d160bd" containerName="barbican-db-sync" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.760478 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2cfadc6-5078-4554-849b-c013b8fa0a2e" containerName="dnsmasq-dns" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.760494 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="42faa16a-d5df-4002-94bc-dd9b727ad202" containerName="placement-db-sync" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.760509 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="858eb7c9-030e-49d8-b018-fdf41e979d20" containerName="keystone-bootstrap" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.761335 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.768958 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r2n9r" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.768970 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.769159 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.769168 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.769228 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.769387 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.770286 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7f958c4db4-6hhpg"] Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.852538 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-public-tls-certs\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.852587 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-internal-tls-certs\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.852623 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-credential-keys\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.852658 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-fernet-keys\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.852682 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-combined-ca-bundle\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.852712 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-scripts\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.852854 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkbl6\" (UniqueName: \"kubernetes.io/projected/d0f0a9aa-0b20-447b-9a08-9d5525010c40-kube-api-access-fkbl6\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.852957 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-config-data\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.883632 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d6f884c6d-xlmbh"] Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.885245 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.889810 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.889858 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.889899 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.889816 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rv2r2" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.890172 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.911242 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d6f884c6d-xlmbh"] Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.924817 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-74556c64c7-5cqch"] Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.926509 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.939682 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-74556c64c7-5cqch"] Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.942043 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-bdtcv" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.942251 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.949575 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954193 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkbl6\" (UniqueName: \"kubernetes.io/projected/d0f0a9aa-0b20-447b-9a08-9d5525010c40-kube-api-access-fkbl6\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954248 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-config-data\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954289 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-public-tls-certs\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954346 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-scripts\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954370 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-internal-tls-certs\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954402 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-public-tls-certs\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954428 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-combined-ca-bundle\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954451 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-internal-tls-certs\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954484 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-credential-keys\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954501 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-config-data\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954530 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-logs\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.954549 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-fernet-keys\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.969243 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-internal-tls-certs\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.969427 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-public-tls-certs\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.969870 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-config-data\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.971952 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fzrq8" event={"ID":"7348b48b-1f84-439f-a049-bfc19935dbfc","Type":"ContainerStarted","Data":"38df6d787c8345ee1050f3c2b8490298e447001c5be7a4220ce71bc14a8df2e0"} Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.974986 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwq89\" (UniqueName: \"kubernetes.io/projected/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-kube-api-access-lwq89\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.975035 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-combined-ca-bundle\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.975081 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-scripts\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:19 crc kubenswrapper[4822]: I1124 14:38:19.977905 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-credential-keys\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.006106 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-fernet-keys\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.013491 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-scripts\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.049322 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkbl6\" (UniqueName: \"kubernetes.io/projected/d0f0a9aa-0b20-447b-9a08-9d5525010c40-kube-api-access-fkbl6\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.049832 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f0a9aa-0b20-447b-9a08-9d5525010c40-combined-ca-bundle\") pod \"keystone-7f958c4db4-6hhpg\" (UID: \"d0f0a9aa-0b20-447b-9a08-9d5525010c40\") " pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.074339 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-75cb666446-dhnfk"] Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.088564 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.090029 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091318 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-scripts\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091354 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-internal-tls-certs\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091383 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75962c4e-af47-4216-99b1-162e00816a30-config-data\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091400 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75962c4e-af47-4216-99b1-162e00816a30-combined-ca-bundle\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091422 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75962c4e-af47-4216-99b1-162e00816a30-config-data-custom\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091449 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-combined-ca-bundle\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091495 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-config-data\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091521 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-logs\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091560 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwq89\" (UniqueName: \"kubernetes.io/projected/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-kube-api-access-lwq89\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091592 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75962c4e-af47-4216-99b1-162e00816a30-logs\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091645 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-public-tls-certs\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.091684 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vscbr\" (UniqueName: \"kubernetes.io/projected/75962c4e-af47-4216-99b1-162e00816a30-kube-api-access-vscbr\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.095806 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.112463 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-logs\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.117886 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-config-data\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.119696 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-scripts\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.121742 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-combined-ca-bundle\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.122161 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75cb666446-dhnfk"] Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.123792 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-public-tls-certs\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.124754 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-internal-tls-certs\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.135408 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwq89\" (UniqueName: \"kubernetes.io/projected/0eb98c2c-948c-4ed6-99b6-f6b352d6d50c-kube-api-access-lwq89\") pod \"placement-d6f884c6d-xlmbh\" (UID: \"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c\") " pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.150916 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-fzrq8" podStartSLOduration=3.490171517 podStartE2EDuration="46.150898592s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="2025-11-24 14:37:35.825326913 +0000 UTC m=+1092.941967390" lastFinishedPulling="2025-11-24 14:38:18.486053968 +0000 UTC m=+1135.602694465" observedRunningTime="2025-11-24 14:38:20.027653627 +0000 UTC m=+1137.144294114" watchObservedRunningTime="2025-11-24 14:38:20.150898592 +0000 UTC m=+1137.267539069" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.173035 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d649d8c65-ptdtg"] Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.178793 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193514 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9242m\" (UniqueName: \"kubernetes.io/projected/045d67b1-30d7-432f-a37e-165ca4eef3b4-kube-api-access-9242m\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193582 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75962c4e-af47-4216-99b1-162e00816a30-logs\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193620 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/045d67b1-30d7-432f-a37e-165ca4eef3b4-logs\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193643 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/045d67b1-30d7-432f-a37e-165ca4eef3b4-config-data\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193704 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/045d67b1-30d7-432f-a37e-165ca4eef3b4-config-data-custom\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193754 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vscbr\" (UniqueName: \"kubernetes.io/projected/75962c4e-af47-4216-99b1-162e00816a30-kube-api-access-vscbr\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193794 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75962c4e-af47-4216-99b1-162e00816a30-config-data\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193811 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75962c4e-af47-4216-99b1-162e00816a30-combined-ca-bundle\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193831 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75962c4e-af47-4216-99b1-162e00816a30-config-data-custom\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.193911 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/045d67b1-30d7-432f-a37e-165ca4eef3b4-combined-ca-bundle\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.194403 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75962c4e-af47-4216-99b1-162e00816a30-logs\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.204339 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d649d8c65-ptdtg"] Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.204802 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75962c4e-af47-4216-99b1-162e00816a30-combined-ca-bundle\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.205299 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.209441 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75962c4e-af47-4216-99b1-162e00816a30-config-data\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.210879 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75962c4e-af47-4216-99b1-162e00816a30-config-data-custom\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.222872 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vscbr\" (UniqueName: \"kubernetes.io/projected/75962c4e-af47-4216-99b1-162e00816a30-kube-api-access-vscbr\") pod \"barbican-worker-74556c64c7-5cqch\" (UID: \"75962c4e-af47-4216-99b1-162e00816a30\") " pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.240708 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-74556c64c7-5cqch" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.248261 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c9b79f5f4-m48xx"] Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.265616 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c9b79f5f4-m48xx"] Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.265720 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.273738 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.295966 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/045d67b1-30d7-432f-a37e-165ca4eef3b4-config-data\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296050 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-dns-svc\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296118 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/045d67b1-30d7-432f-a37e-165ca4eef3b4-config-data-custom\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296218 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-sb\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296319 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljqsg\" (UniqueName: \"kubernetes.io/projected/fd67711e-c7f9-4cc8-85e8-1886545615f2-kube-api-access-ljqsg\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296345 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-config\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296388 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/045d67b1-30d7-432f-a37e-165ca4eef3b4-combined-ca-bundle\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296427 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-nb\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296454 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9242m\" (UniqueName: \"kubernetes.io/projected/045d67b1-30d7-432f-a37e-165ca4eef3b4-kube-api-access-9242m\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296502 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/045d67b1-30d7-432f-a37e-165ca4eef3b4-logs\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.296952 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/045d67b1-30d7-432f-a37e-165ca4eef3b4-logs\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.310882 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/045d67b1-30d7-432f-a37e-165ca4eef3b4-config-data-custom\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.313325 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/045d67b1-30d7-432f-a37e-165ca4eef3b4-config-data\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.313972 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/045d67b1-30d7-432f-a37e-165ca4eef3b4-combined-ca-bundle\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.315925 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9242m\" (UniqueName: \"kubernetes.io/projected/045d67b1-30d7-432f-a37e-165ca4eef3b4-kube-api-access-9242m\") pod \"barbican-keystone-listener-75cb666446-dhnfk\" (UID: \"045d67b1-30d7-432f-a37e-165ca4eef3b4\") " pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409257 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnhpp\" (UniqueName: \"kubernetes.io/projected/23b52820-32f7-4734-a60d-c7d0b21206b5-kube-api-access-vnhpp\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409358 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-dns-svc\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409378 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data-custom\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409431 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-sb\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409467 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljqsg\" (UniqueName: \"kubernetes.io/projected/fd67711e-c7f9-4cc8-85e8-1886545615f2-kube-api-access-ljqsg\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409490 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-config\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409508 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-combined-ca-bundle\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409562 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409582 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b52820-32f7-4734-a60d-c7d0b21206b5-logs\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.409602 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-nb\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.415124 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-sb\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.415634 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-config\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.416448 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-dns-svc\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.417434 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-nb\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.459002 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljqsg\" (UniqueName: \"kubernetes.io/projected/fd67711e-c7f9-4cc8-85e8-1886545615f2-kube-api-access-ljqsg\") pod \"dnsmasq-dns-7d649d8c65-ptdtg\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.532278 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b52820-32f7-4734-a60d-c7d0b21206b5-logs\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.532365 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnhpp\" (UniqueName: \"kubernetes.io/projected/23b52820-32f7-4734-a60d-c7d0b21206b5-kube-api-access-vnhpp\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.532438 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data-custom\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.532521 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-combined-ca-bundle\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.532579 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.535633 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b52820-32f7-4734-a60d-c7d0b21206b5-logs\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.539507 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.539986 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data-custom\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.556500 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.556718 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-combined-ca-bundle\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.564001 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnhpp\" (UniqueName: \"kubernetes.io/projected/23b52820-32f7-4734-a60d-c7d0b21206b5-kube-api-access-vnhpp\") pod \"barbican-api-6c9b79f5f4-m48xx\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.574877 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.629723 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.671259 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.671304 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.693756 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7f958c4db4-6hhpg"] Nov 24 14:38:20 crc kubenswrapper[4822]: W1124 14:38:20.725926 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0f0a9aa_0b20_447b_9a08_9d5525010c40.slice/crio-b5cf05f1418ae2e784cd7ab9d05ccbb1b90abcc98ef27c659840ab00751dd375 WatchSource:0}: Error finding container b5cf05f1418ae2e784cd7ab9d05ccbb1b90abcc98ef27c659840ab00751dd375: Status 404 returned error can't find the container with id b5cf05f1418ae2e784cd7ab9d05ccbb1b90abcc98ef27c659840ab00751dd375 Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.729291 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.806834 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.943522 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d6f884c6d-xlmbh"] Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.995556 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f958c4db4-6hhpg" event={"ID":"d0f0a9aa-0b20-447b-9a08-9d5525010c40","Type":"ContainerStarted","Data":"b5cf05f1418ae2e784cd7ab9d05ccbb1b90abcc98ef27c659840ab00751dd375"} Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.995898 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:38:20 crc kubenswrapper[4822]: I1124 14:38:20.996440 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:38:21 crc kubenswrapper[4822]: I1124 14:38:21.039259 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-74556c64c7-5cqch"] Nov 24 14:38:21 crc kubenswrapper[4822]: W1124 14:38:21.051131 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75962c4e_af47_4216_99b1_162e00816a30.slice/crio-7daa7303f90799cd5ae1030bfeabcb352df7fd4d3373d5bdea9a2c1e7b2c39e9 WatchSource:0}: Error finding container 7daa7303f90799cd5ae1030bfeabcb352df7fd4d3373d5bdea9a2c1e7b2c39e9: Status 404 returned error can't find the container with id 7daa7303f90799cd5ae1030bfeabcb352df7fd4d3373d5bdea9a2c1e7b2c39e9 Nov 24 14:38:21 crc kubenswrapper[4822]: I1124 14:38:21.207309 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75cb666446-dhnfk"] Nov 24 14:38:21 crc kubenswrapper[4822]: I1124 14:38:21.216751 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d649d8c65-ptdtg"] Nov 24 14:38:21 crc kubenswrapper[4822]: I1124 14:38:21.477114 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c9b79f5f4-m48xx"] Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.014940 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d6f884c6d-xlmbh" event={"ID":"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c","Type":"ContainerStarted","Data":"721c89abe5a108f4582066f4a603154466a373de44d137bfecca8b49d9d32aff"} Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.015193 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d6f884c6d-xlmbh" event={"ID":"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c","Type":"ContainerStarted","Data":"9f7b9b61233411649b73a8eca8107893578780f73cd8a68297a91db514232cd8"} Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.018358 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" event={"ID":"045d67b1-30d7-432f-a37e-165ca4eef3b4","Type":"ContainerStarted","Data":"0ce755e1a35f0f9ce267bb4f2fd403d24b22433537c1bd4be2ed605b6fa8fc82"} Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.020003 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-74556c64c7-5cqch" event={"ID":"75962c4e-af47-4216-99b1-162e00816a30","Type":"ContainerStarted","Data":"7daa7303f90799cd5ae1030bfeabcb352df7fd4d3373d5bdea9a2c1e7b2c39e9"} Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.033629 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9b79f5f4-m48xx" event={"ID":"23b52820-32f7-4734-a60d-c7d0b21206b5","Type":"ContainerStarted","Data":"e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc"} Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.033676 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9b79f5f4-m48xx" event={"ID":"23b52820-32f7-4734-a60d-c7d0b21206b5","Type":"ContainerStarted","Data":"9f202d44f968d4db6fb912f3643e1db192ddc3e6c485c55c57e287d08d9efbd0"} Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.038976 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f958c4db4-6hhpg" event={"ID":"d0f0a9aa-0b20-447b-9a08-9d5525010c40","Type":"ContainerStarted","Data":"25861a473bed5cf0930ec3e05afe85b0bdba0d38f615daedeb15709eae294ca0"} Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.040073 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.044812 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" event={"ID":"fd67711e-c7f9-4cc8-85e8-1886545615f2","Type":"ContainerStarted","Data":"0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24"} Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.044859 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" event={"ID":"fd67711e-c7f9-4cc8-85e8-1886545615f2","Type":"ContainerStarted","Data":"ad397f36e0e8906132cbd799780465c54f6bb81c60591cc8b30fc832ae81d551"} Nov 24 14:38:22 crc kubenswrapper[4822]: I1124 14:38:22.058098 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7f958c4db4-6hhpg" podStartSLOduration=3.058083619 podStartE2EDuration="3.058083619s" podCreationTimestamp="2025-11-24 14:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:22.057555713 +0000 UTC m=+1139.174196190" watchObservedRunningTime="2025-11-24 14:38:22.058083619 +0000 UTC m=+1139.174724096" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.056280 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9b79f5f4-m48xx" event={"ID":"23b52820-32f7-4734-a60d-c7d0b21206b5","Type":"ContainerStarted","Data":"379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b"} Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.056935 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.056949 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.059027 4822 generic.go:334] "Generic (PLEG): container finished" podID="fd67711e-c7f9-4cc8-85e8-1886545615f2" containerID="0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24" exitCode=0 Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.059090 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" event={"ID":"fd67711e-c7f9-4cc8-85e8-1886545615f2","Type":"ContainerDied","Data":"0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24"} Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.065353 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d6f884c6d-xlmbh" event={"ID":"0eb98c2c-948c-4ed6-99b6-f6b352d6d50c","Type":"ContainerStarted","Data":"9018e7aba7eb9f1ae0196aa0f4448966df41e3d7eeaad59bfed0feeb2f68c80b"} Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.065373 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.065420 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.065715 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.065741 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.096239 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c9b79f5f4-m48xx" podStartSLOduration=3.096219043 podStartE2EDuration="3.096219043s" podCreationTimestamp="2025-11-24 14:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:23.077193566 +0000 UTC m=+1140.193834053" watchObservedRunningTime="2025-11-24 14:38:23.096219043 +0000 UTC m=+1140.212859520" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.110687 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d6f884c6d-xlmbh" podStartSLOduration=4.110667089 podStartE2EDuration="4.110667089s" podCreationTimestamp="2025-11-24 14:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:23.100699492 +0000 UTC m=+1140.217339969" watchObservedRunningTime="2025-11-24 14:38:23.110667089 +0000 UTC m=+1140.227307566" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.545665 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:38:23 crc kubenswrapper[4822]: I1124 14:38:23.795763 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.028443 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5cbff998cd-sdjjk"] Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.038025 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.046444 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.046606 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.067262 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5cbff998cd-sdjjk"] Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.157019 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-public-tls-certs\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.157062 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-config-data\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.157238 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq5sx\" (UniqueName: \"kubernetes.io/projected/22ebfa19-b296-49c5-888b-cadf1e18d897-kube-api-access-vq5sx\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.157425 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-config-data-custom\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.157495 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-combined-ca-bundle\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.157536 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22ebfa19-b296-49c5-888b-cadf1e18d897-logs\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.157667 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-internal-tls-certs\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.203557 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.203690 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.259299 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-config-data\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.259398 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq5sx\" (UniqueName: \"kubernetes.io/projected/22ebfa19-b296-49c5-888b-cadf1e18d897-kube-api-access-vq5sx\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.259507 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-config-data-custom\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.259560 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-combined-ca-bundle\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.259616 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22ebfa19-b296-49c5-888b-cadf1e18d897-logs\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.259635 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-internal-tls-certs\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.259761 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-public-tls-certs\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.262381 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22ebfa19-b296-49c5-888b-cadf1e18d897-logs\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.265606 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-internal-tls-certs\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.265945 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-config-data-custom\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.266821 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-config-data\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.270039 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-public-tls-certs\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.280224 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq5sx\" (UniqueName: \"kubernetes.io/projected/22ebfa19-b296-49c5-888b-cadf1e18d897-kube-api-access-vq5sx\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.281015 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ebfa19-b296-49c5-888b-cadf1e18d897-combined-ca-bundle\") pod \"barbican-api-5cbff998cd-sdjjk\" (UID: \"22ebfa19-b296-49c5-888b-cadf1e18d897\") " pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.385126 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:24 crc kubenswrapper[4822]: I1124 14:38:24.627349 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:38:25 crc kubenswrapper[4822]: I1124 14:38:25.110358 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-74556c64c7-5cqch" event={"ID":"75962c4e-af47-4216-99b1-162e00816a30","Type":"ContainerStarted","Data":"4761f92f2c9d140aad0586630247c7de4656560abd0172f1baa71a47350b1e60"} Nov 24 14:38:25 crc kubenswrapper[4822]: I1124 14:38:25.113074 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-s6qvf" event={"ID":"f0ad3a91-d60c-40a0-a285-c5a9615fffe0","Type":"ContainerStarted","Data":"25897687c5bab918dbe59432ac3f4cc18ba2f64eaff93cc50236c796cd596630"} Nov 24 14:38:25 crc kubenswrapper[4822]: I1124 14:38:25.116143 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" event={"ID":"045d67b1-30d7-432f-a37e-165ca4eef3b4","Type":"ContainerStarted","Data":"6256bef1ab4c47fb890666da13d96a6cb137cd82d67e8000aad175e505bbaad1"} Nov 24 14:38:25 crc kubenswrapper[4822]: I1124 14:38:25.121585 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" event={"ID":"fd67711e-c7f9-4cc8-85e8-1886545615f2","Type":"ContainerStarted","Data":"610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244"} Nov 24 14:38:25 crc kubenswrapper[4822]: I1124 14:38:25.162654 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-s6qvf" podStartSLOduration=2.638910916 podStartE2EDuration="51.162633456s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="2025-11-24 14:37:36.089373838 +0000 UTC m=+1093.206014315" lastFinishedPulling="2025-11-24 14:38:24.613096388 +0000 UTC m=+1141.729736855" observedRunningTime="2025-11-24 14:38:25.132464975 +0000 UTC m=+1142.249105462" watchObservedRunningTime="2025-11-24 14:38:25.162633456 +0000 UTC m=+1142.279273933" Nov 24 14:38:25 crc kubenswrapper[4822]: I1124 14:38:25.165445 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5cbff998cd-sdjjk"] Nov 24 14:38:25 crc kubenswrapper[4822]: I1124 14:38:25.169623 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" podStartSLOduration=5.169614132 podStartE2EDuration="5.169614132s" podCreationTimestamp="2025-11-24 14:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:25.157642362 +0000 UTC m=+1142.274282849" watchObservedRunningTime="2025-11-24 14:38:25.169614132 +0000 UTC m=+1142.286254609" Nov 24 14:38:25 crc kubenswrapper[4822]: I1124 14:38:25.575536 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.148427 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cbff998cd-sdjjk" event={"ID":"22ebfa19-b296-49c5-888b-cadf1e18d897","Type":"ContainerStarted","Data":"a386777f8ad46f3eed48422e6f6ca9810c70719021b03641e1153db421cf139f"} Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.148782 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cbff998cd-sdjjk" event={"ID":"22ebfa19-b296-49c5-888b-cadf1e18d897","Type":"ContainerStarted","Data":"2c37a0f88a07443ae253d18866814f97a6e37b8dee178975e9805802b4d34d3b"} Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.148804 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cbff998cd-sdjjk" event={"ID":"22ebfa19-b296-49c5-888b-cadf1e18d897","Type":"ContainerStarted","Data":"7549976e9dcb56fc51ea7bb3c5ecfddbf06a3c77139192a9ac52042991f7d751"} Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.148876 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.148953 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.153547 4822 generic.go:334] "Generic (PLEG): container finished" podID="7348b48b-1f84-439f-a049-bfc19935dbfc" containerID="38df6d787c8345ee1050f3c2b8490298e447001c5be7a4220ce71bc14a8df2e0" exitCode=0 Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.153612 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fzrq8" event={"ID":"7348b48b-1f84-439f-a049-bfc19935dbfc","Type":"ContainerDied","Data":"38df6d787c8345ee1050f3c2b8490298e447001c5be7a4220ce71bc14a8df2e0"} Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.159578 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" event={"ID":"045d67b1-30d7-432f-a37e-165ca4eef3b4","Type":"ContainerStarted","Data":"c33a4fb470e63f8a04973e77b51447ded104c79538d61a5a64c43dea9c589db5"} Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.163626 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-74556c64c7-5cqch" event={"ID":"75962c4e-af47-4216-99b1-162e00816a30","Type":"ContainerStarted","Data":"1bf0f9799c81b129d4c6d12876e4bd7985f21b5786f223acdef30cfe31498cd0"} Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.189062 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5cbff998cd-sdjjk" podStartSLOduration=3.189042367 podStartE2EDuration="3.189042367s" podCreationTimestamp="2025-11-24 14:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:26.175147008 +0000 UTC m=+1143.291787525" watchObservedRunningTime="2025-11-24 14:38:26.189042367 +0000 UTC m=+1143.305682854" Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.208921 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-75cb666446-dhnfk" podStartSLOduration=3.82731236 podStartE2EDuration="7.208900211s" podCreationTimestamp="2025-11-24 14:38:19 +0000 UTC" firstStartedPulling="2025-11-24 14:38:21.238968588 +0000 UTC m=+1138.355609065" lastFinishedPulling="2025-11-24 14:38:24.620556439 +0000 UTC m=+1141.737196916" observedRunningTime="2025-11-24 14:38:26.201968657 +0000 UTC m=+1143.318609144" watchObservedRunningTime="2025-11-24 14:38:26.208900211 +0000 UTC m=+1143.325540688" Nov 24 14:38:26 crc kubenswrapper[4822]: I1124 14:38:26.243633 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-74556c64c7-5cqch" podStartSLOduration=3.685171842 podStartE2EDuration="7.243616583s" podCreationTimestamp="2025-11-24 14:38:19 +0000 UTC" firstStartedPulling="2025-11-24 14:38:21.054244305 +0000 UTC m=+1138.170884782" lastFinishedPulling="2025-11-24 14:38:24.612689046 +0000 UTC m=+1141.729329523" observedRunningTime="2025-11-24 14:38:26.242841919 +0000 UTC m=+1143.359482416" watchObservedRunningTime="2025-11-24 14:38:26.243616583 +0000 UTC m=+1143.360257060" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.139269 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.206338 4822 generic.go:334] "Generic (PLEG): container finished" podID="f0ad3a91-d60c-40a0-a285-c5a9615fffe0" containerID="25897687c5bab918dbe59432ac3f4cc18ba2f64eaff93cc50236c796cd596630" exitCode=0 Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.206411 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-s6qvf" event={"ID":"f0ad3a91-d60c-40a0-a285-c5a9615fffe0","Type":"ContainerDied","Data":"25897687c5bab918dbe59432ac3f4cc18ba2f64eaff93cc50236c796cd596630"} Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.208854 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fzrq8" event={"ID":"7348b48b-1f84-439f-a049-bfc19935dbfc","Type":"ContainerDied","Data":"b496b8bc17235a62e228ad291c7b1ab2a3bcdaead4c0a12b56811dae29e13346"} Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.208885 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fzrq8" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.208896 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b496b8bc17235a62e228ad291c7b1ab2a3bcdaead4c0a12b56811dae29e13346" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.215334 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-config-data\") pod \"7348b48b-1f84-439f-a049-bfc19935dbfc\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.215415 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7348b48b-1f84-439f-a049-bfc19935dbfc-etc-machine-id\") pod \"7348b48b-1f84-439f-a049-bfc19935dbfc\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.215522 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-db-sync-config-data\") pod \"7348b48b-1f84-439f-a049-bfc19935dbfc\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.215561 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-combined-ca-bundle\") pod \"7348b48b-1f84-439f-a049-bfc19935dbfc\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.215603 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r72z7\" (UniqueName: \"kubernetes.io/projected/7348b48b-1f84-439f-a049-bfc19935dbfc-kube-api-access-r72z7\") pod \"7348b48b-1f84-439f-a049-bfc19935dbfc\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.215830 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-scripts\") pod \"7348b48b-1f84-439f-a049-bfc19935dbfc\" (UID: \"7348b48b-1f84-439f-a049-bfc19935dbfc\") " Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.228319 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7348b48b-1f84-439f-a049-bfc19935dbfc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7348b48b-1f84-439f-a049-bfc19935dbfc" (UID: "7348b48b-1f84-439f-a049-bfc19935dbfc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.241274 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7348b48b-1f84-439f-a049-bfc19935dbfc" (UID: "7348b48b-1f84-439f-a049-bfc19935dbfc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.252517 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7348b48b-1f84-439f-a049-bfc19935dbfc-kube-api-access-r72z7" (OuterVolumeSpecName: "kube-api-access-r72z7") pod "7348b48b-1f84-439f-a049-bfc19935dbfc" (UID: "7348b48b-1f84-439f-a049-bfc19935dbfc"). InnerVolumeSpecName "kube-api-access-r72z7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.252713 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-scripts" (OuterVolumeSpecName: "scripts") pod "7348b48b-1f84-439f-a049-bfc19935dbfc" (UID: "7348b48b-1f84-439f-a049-bfc19935dbfc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.305397 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7348b48b-1f84-439f-a049-bfc19935dbfc" (UID: "7348b48b-1f84-439f-a049-bfc19935dbfc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.328747 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.328796 4822 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7348b48b-1f84-439f-a049-bfc19935dbfc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.328809 4822 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.328818 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.328828 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r72z7\" (UniqueName: \"kubernetes.io/projected/7348b48b-1f84-439f-a049-bfc19935dbfc-kube-api-access-r72z7\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.406414 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-config-data" (OuterVolumeSpecName: "config-data") pod "7348b48b-1f84-439f-a049-bfc19935dbfc" (UID: "7348b48b-1f84-439f-a049-bfc19935dbfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.430322 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7348b48b-1f84-439f-a049-bfc19935dbfc-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.577172 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.628945 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fb745b69-962h6"] Nov 24 14:38:30 crc kubenswrapper[4822]: I1124 14:38:30.629189 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fb745b69-962h6" podUID="3f0b9893-50bf-4550-a893-bd82d7e4ecc3" containerName="dnsmasq-dns" containerID="cri-o://a6d0ce151b87bdef67f37e6d46b107fd8db4222d090cef61bea25935634d62ab" gracePeriod=10 Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.218985 4822 generic.go:334] "Generic (PLEG): container finished" podID="3f0b9893-50bf-4550-a893-bd82d7e4ecc3" containerID="a6d0ce151b87bdef67f37e6d46b107fd8db4222d090cef61bea25935634d62ab" exitCode=0 Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.219304 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fb745b69-962h6" event={"ID":"3f0b9893-50bf-4550-a893-bd82d7e4ecc3","Type":"ContainerDied","Data":"a6d0ce151b87bdef67f37e6d46b107fd8db4222d090cef61bea25935634d62ab"} Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.620407 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:38:31 crc kubenswrapper[4822]: E1124 14:38:31.621137 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7348b48b-1f84-439f-a049-bfc19935dbfc" containerName="cinder-db-sync" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.621150 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="7348b48b-1f84-439f-a049-bfc19935dbfc" containerName="cinder-db-sync" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.621361 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="7348b48b-1f84-439f-a049-bfc19935dbfc" containerName="cinder-db-sync" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.622639 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.627819 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.628019 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.628589 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.628699 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-fzfh7" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.643079 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57fff66767-fnp4z"] Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.644841 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.679383 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.709933 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.748383 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57fff66767-fnp4z"] Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800174 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2vft\" (UniqueName: \"kubernetes.io/projected/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-kube-api-access-j2vft\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800256 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800280 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-scripts\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800303 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800337 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-nb\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800362 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800384 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-config\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800406 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-sb\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800425 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800443 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smnpf\" (UniqueName: \"kubernetes.io/projected/8610fa72-47f6-4d01-be06-b1d3486c398b-kube-api-access-smnpf\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.800515 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-dns-svc\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.811586 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.811662 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:38:31 crc kubenswrapper[4822]: E1124 14:38:31.812392 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f0b9893-50bf-4550-a893-bd82d7e4ecc3" containerName="dnsmasq-dns" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.812406 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f0b9893-50bf-4550-a893-bd82d7e4ecc3" containerName="dnsmasq-dns" Nov 24 14:38:31 crc kubenswrapper[4822]: E1124 14:38:31.812423 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ad3a91-d60c-40a0-a285-c5a9615fffe0" containerName="cloudkitty-db-sync" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.812430 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ad3a91-d60c-40a0-a285-c5a9615fffe0" containerName="cloudkitty-db-sync" Nov 24 14:38:31 crc kubenswrapper[4822]: E1124 14:38:31.812464 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f0b9893-50bf-4550-a893-bd82d7e4ecc3" containerName="init" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.812470 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f0b9893-50bf-4550-a893-bd82d7e4ecc3" containerName="init" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.812650 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ad3a91-d60c-40a0-a285-c5a9615fffe0" containerName="cloudkitty-db-sync" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.812665 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f0b9893-50bf-4550-a893-bd82d7e4ecc3" containerName="dnsmasq-dns" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.827719 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.827829 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.830220 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.903072 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-dns-svc\") pod \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.904054 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-nb\") pod \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.904127 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-sb\") pod \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.904162 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-scripts\") pod \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.904240 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkqwt\" (UniqueName: \"kubernetes.io/projected/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-kube-api-access-nkqwt\") pod \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.904264 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-config-data\") pod \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.904291 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-config\") pod \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.911561 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-dns-svc\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.911679 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2vft\" (UniqueName: \"kubernetes.io/projected/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-kube-api-access-j2vft\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.911779 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.911798 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-scripts\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.911841 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.911902 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-nb\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.911982 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.912015 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-config\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.912074 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-sb\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.912104 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.912378 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smnpf\" (UniqueName: \"kubernetes.io/projected/8610fa72-47f6-4d01-be06-b1d3486c398b-kube-api-access-smnpf\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.912472 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-dns-svc\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.912967 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.914604 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-config\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.915196 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-sb\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.922068 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-nb\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.930625 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.931330 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-scripts\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.941568 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smnpf\" (UniqueName: \"kubernetes.io/projected/8610fa72-47f6-4d01-be06-b1d3486c398b-kube-api-access-smnpf\") pod \"dnsmasq-dns-57fff66767-fnp4z\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.945398 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-kube-api-access-nkqwt" (OuterVolumeSpecName: "kube-api-access-nkqwt") pod "3f0b9893-50bf-4550-a893-bd82d7e4ecc3" (UID: "3f0b9893-50bf-4550-a893-bd82d7e4ecc3"). InnerVolumeSpecName "kube-api-access-nkqwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.949346 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-scripts" (OuterVolumeSpecName: "scripts") pod "f0ad3a91-d60c-40a0-a285-c5a9615fffe0" (UID: "f0ad3a91-d60c-40a0-a285-c5a9615fffe0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.949856 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2vft\" (UniqueName: \"kubernetes.io/projected/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-kube-api-access-j2vft\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.960726 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:31 crc kubenswrapper[4822]: I1124 14:38:31.961054 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.015827 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3f0b9893-50bf-4550-a893-bd82d7e4ecc3" (UID: "3f0b9893-50bf-4550-a893-bd82d7e4ecc3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.015900 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-nb\") pod \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\" (UID: \"3f0b9893-50bf-4550-a893-bd82d7e4ecc3\") " Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.015936 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-combined-ca-bundle\") pod \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.015969 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx97z\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-kube-api-access-rx97z\") pod \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.016063 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-certs\") pod \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\" (UID: \"f0ad3a91-d60c-40a0-a285-c5a9615fffe0\") " Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.017178 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data-custom\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.017321 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: W1124 14:38:32.019172 4822 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/3f0b9893-50bf-4550-a893-bd82d7e4ecc3/volumes/kubernetes.io~configmap/ovsdbserver-nb Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.019226 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3f0b9893-50bf-4550-a893-bd82d7e4ecc3" (UID: "3f0b9893-50bf-4550-a893-bd82d7e4ecc3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.029293 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhq64\" (UniqueName: \"kubernetes.io/projected/a0aeb993-7d61-4924-bc31-86c00d8accb5-kube-api-access-hhq64\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.029337 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0aeb993-7d61-4924-bc31-86c00d8accb5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.029387 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.029450 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-kube-api-access-rx97z" (OuterVolumeSpecName: "kube-api-access-rx97z") pod "f0ad3a91-d60c-40a0-a285-c5a9615fffe0" (UID: "f0ad3a91-d60c-40a0-a285-c5a9615fffe0"). InnerVolumeSpecName "kube-api-access-rx97z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.029591 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-certs" (OuterVolumeSpecName: "certs") pod "f0ad3a91-d60c-40a0-a285-c5a9615fffe0" (UID: "f0ad3a91-d60c-40a0-a285-c5a9615fffe0"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.030650 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-scripts\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.030704 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0aeb993-7d61-4924-bc31-86c00d8accb5-logs\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.030844 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.030860 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx97z\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-kube-api-access-rx97z\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.030871 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.030879 4822 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.030889 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkqwt\" (UniqueName: \"kubernetes.io/projected/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-kube-api-access-nkqwt\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.040548 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.046706 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.053323 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-config-data" (OuterVolumeSpecName: "config-data") pod "f0ad3a91-d60c-40a0-a285-c5a9615fffe0" (UID: "f0ad3a91-d60c-40a0-a285-c5a9615fffe0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.074909 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3f0b9893-50bf-4550-a893-bd82d7e4ecc3" (UID: "3f0b9893-50bf-4550-a893-bd82d7e4ecc3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.106822 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3f0b9893-50bf-4550-a893-bd82d7e4ecc3" (UID: "3f0b9893-50bf-4550-a893-bd82d7e4ecc3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.109014 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0ad3a91-d60c-40a0-a285-c5a9615fffe0" (UID: "f0ad3a91-d60c-40a0-a285-c5a9615fffe0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.112218 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-config" (OuterVolumeSpecName: "config") pod "3f0b9893-50bf-4550-a893-bd82d7e4ecc3" (UID: "3f0b9893-50bf-4550-a893-bd82d7e4ecc3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132331 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132420 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-scripts\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132441 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0aeb993-7d61-4924-bc31-86c00d8accb5-logs\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132542 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data-custom\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132571 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132612 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhq64\" (UniqueName: \"kubernetes.io/projected/a0aeb993-7d61-4924-bc31-86c00d8accb5-kube-api-access-hhq64\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132630 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0aeb993-7d61-4924-bc31-86c00d8accb5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132680 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132692 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132703 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132712 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ad3a91-d60c-40a0-a285-c5a9615fffe0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132722 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0b9893-50bf-4550-a893-bd82d7e4ecc3-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.132761 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0aeb993-7d61-4924-bc31-86c00d8accb5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.134791 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0aeb993-7d61-4924-bc31-86c00d8accb5-logs\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.143121 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.177445 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-scripts\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.177975 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data-custom\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.178807 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.198841 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhq64\" (UniqueName: \"kubernetes.io/projected/a0aeb993-7d61-4924-bc31-86c00d8accb5-kube-api-access-hhq64\") pod \"cinder-api-0\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.250255 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fb745b69-962h6" event={"ID":"3f0b9893-50bf-4550-a893-bd82d7e4ecc3","Type":"ContainerDied","Data":"bfbd4d0e928c869f7dd1d3c71fb58a75ce9775d395299d17582d338e34cef78a"} Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.250325 4822 scope.go:117] "RemoveContainer" containerID="a6d0ce151b87bdef67f37e6d46b107fd8db4222d090cef61bea25935634d62ab" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.250510 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fb745b69-962h6" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.300138 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-s6qvf" event={"ID":"f0ad3a91-d60c-40a0-a285-c5a9615fffe0","Type":"ContainerDied","Data":"a6e070c2971b0b872f6258f250bdc1be079cab6033f4aaee409f90bf978a4c9c"} Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.300185 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6e070c2971b0b872f6258f250bdc1be079cab6033f4aaee409f90bf978a4c9c" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.300268 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-s6qvf" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.368161 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fb745b69-962h6"] Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.395588 4822 scope.go:117] "RemoveContainer" containerID="d9f4f5ac8fdfc931f03470a4d5f62861c24ce40b5b9df4e438cf6a21a4b1393c" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.404271 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fb745b69-962h6"] Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.451749 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-j72lq"] Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.453374 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.458091 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.458278 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.458575 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.458601 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.458710 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-zhp7h" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.489557 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-j72lq"] Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.490333 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.556854 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-combined-ca-bundle\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.556916 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-scripts\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.556941 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8v2d\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-kube-api-access-t8v2d\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.556985 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-config-data\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.557004 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-certs\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: E1124 14:38:32.578489 4822 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f0b9893_50bf_4550_a893_bd82d7e4ecc3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f0b9893_50bf_4550_a893_bd82d7e4ecc3.slice/crio-bfbd4d0e928c869f7dd1d3c71fb58a75ce9775d395299d17582d338e34cef78a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0ad3a91_d60c_40a0_a285_c5a9615fffe0.slice/crio-a6e070c2971b0b872f6258f250bdc1be079cab6033f4aaee409f90bf978a4c9c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0ad3a91_d60c_40a0_a285_c5a9615fffe0.slice\": RecentStats: unable to find data in memory cache]" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.662345 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-config-data\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.662396 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-certs\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.662518 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-combined-ca-bundle\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.662564 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-scripts\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.662587 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8v2d\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-kube-api-access-t8v2d\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.673902 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-certs\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.673970 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-combined-ca-bundle\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.675689 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-scripts\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.690476 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8v2d\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-kube-api-access-t8v2d\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.690739 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-config-data\") pod \"cloudkitty-storageinit-j72lq\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.791896 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57fff66767-fnp4z"] Nov 24 14:38:32 crc kubenswrapper[4822]: W1124 14:38:32.796451 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8610fa72_47f6_4d01_be06_b1d3486c398b.slice/crio-ab3996d3ec43ea51e826579d69c252cdb3a909048d3af274bf5b5c57f1c98752 WatchSource:0}: Error finding container ab3996d3ec43ea51e826579d69c252cdb3a909048d3af274bf5b5c57f1c98752: Status 404 returned error can't find the container with id ab3996d3ec43ea51e826579d69c252cdb3a909048d3af274bf5b5c57f1c98752 Nov 24 14:38:32 crc kubenswrapper[4822]: I1124 14:38:32.830676 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.079724 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.201634 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.223182 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.295669 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.333832 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.352996 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0fbdcd8a-a688-4c39-b968-36c5d3abecbd","Type":"ContainerStarted","Data":"4ea03ac80448c094bcb137c517c41f8967174f22ebb994fc6266ab43215094a4"} Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.354380 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a0aeb993-7d61-4924-bc31-86c00d8accb5","Type":"ContainerStarted","Data":"efff0764f6c6d93fc8d1fa74d29a690714957ae90d21d29e67c4c068aa553df7"} Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.356053 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerStarted","Data":"fcffa52afa71bb22a16f640ca7bc1595c95e0c1623ceda38c4d474445aa8eb81"} Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.356280 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="ceilometer-central-agent" containerID="cri-o://f9a32f0cc45408679a748bfb9c5fe147c7e9c92bf6130f47e7e57fe7e0b2eaa7" gracePeriod=30 Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.356570 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.356685 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="sg-core" containerID="cri-o://6fd3b5a8ca9cfaa068d4131029730208157f302f41fbcd2df94c39285f3a9420" gracePeriod=30 Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.356855 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="proxy-httpd" containerID="cri-o://fcffa52afa71bb22a16f640ca7bc1595c95e0c1623ceda38c4d474445aa8eb81" gracePeriod=30 Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.356965 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="ceilometer-notification-agent" containerID="cri-o://13e53dd3c7375c0c69232d41aa2e75146658d67590f39698fddd3dc82d27f7a8" gracePeriod=30 Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.392708 4822 generic.go:334] "Generic (PLEG): container finished" podID="8610fa72-47f6-4d01-be06-b1d3486c398b" containerID="09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2" exitCode=0 Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.393987 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" event={"ID":"8610fa72-47f6-4d01-be06-b1d3486c398b","Type":"ContainerDied","Data":"09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2"} Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.394082 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" event={"ID":"8610fa72-47f6-4d01-be06-b1d3486c398b","Type":"ContainerStarted","Data":"ab3996d3ec43ea51e826579d69c252cdb3a909048d3af274bf5b5c57f1c98752"} Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.425852 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.524164955 podStartE2EDuration="59.425830503s" podCreationTimestamp="2025-11-24 14:37:34 +0000 UTC" firstStartedPulling="2025-11-24 14:37:36.311150455 +0000 UTC m=+1093.427790932" lastFinishedPulling="2025-11-24 14:38:31.212816003 +0000 UTC m=+1148.329456480" observedRunningTime="2025-11-24 14:38:33.411955554 +0000 UTC m=+1150.528596031" watchObservedRunningTime="2025-11-24 14:38:33.425830503 +0000 UTC m=+1150.542470970" Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.437575 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-j72lq"] Nov 24 14:38:33 crc kubenswrapper[4822]: I1124 14:38:33.721920 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f0b9893-50bf-4550-a893-bd82d7e4ecc3" path="/var/lib/kubelet/pods/3f0b9893-50bf-4550-a893-bd82d7e4ecc3/volumes" Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.130966 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.421615 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" event={"ID":"8610fa72-47f6-4d01-be06-b1d3486c398b","Type":"ContainerStarted","Data":"fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8"} Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.422007 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.425608 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a0aeb993-7d61-4924-bc31-86c00d8accb5","Type":"ContainerStarted","Data":"ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c"} Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.433235 4822 generic.go:334] "Generic (PLEG): container finished" podID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerID="fcffa52afa71bb22a16f640ca7bc1595c95e0c1623ceda38c4d474445aa8eb81" exitCode=0 Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.433268 4822 generic.go:334] "Generic (PLEG): container finished" podID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerID="6fd3b5a8ca9cfaa068d4131029730208157f302f41fbcd2df94c39285f3a9420" exitCode=2 Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.433276 4822 generic.go:334] "Generic (PLEG): container finished" podID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerID="f9a32f0cc45408679a748bfb9c5fe147c7e9c92bf6130f47e7e57fe7e0b2eaa7" exitCode=0 Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.433313 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerDied","Data":"fcffa52afa71bb22a16f640ca7bc1595c95e0c1623ceda38c4d474445aa8eb81"} Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.433340 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerDied","Data":"6fd3b5a8ca9cfaa068d4131029730208157f302f41fbcd2df94c39285f3a9420"} Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.433349 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerDied","Data":"f9a32f0cc45408679a748bfb9c5fe147c7e9c92bf6130f47e7e57fe7e0b2eaa7"} Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.436021 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-j72lq" event={"ID":"c25c6942-9f38-4db4-834f-2262ee899ed7","Type":"ContainerStarted","Data":"16409edc9499b818433aed5a90f6f114cc12534d16259515b0d3505fd5a7f229"} Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.436061 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-j72lq" event={"ID":"c25c6942-9f38-4db4-834f-2262ee899ed7","Type":"ContainerStarted","Data":"311ae1113bb8da0e8915ceea0937843d572bf289ce90e06ad719fa1ca6a4edc7"} Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.451790 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" podStartSLOduration=3.45177151 podStartE2EDuration="3.45177151s" podCreationTimestamp="2025-11-24 14:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:34.441328898 +0000 UTC m=+1151.557969395" watchObservedRunningTime="2025-11-24 14:38:34.45177151 +0000 UTC m=+1151.568411987" Nov 24 14:38:34 crc kubenswrapper[4822]: I1124 14:38:34.469912 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-j72lq" podStartSLOduration=2.46989692 podStartE2EDuration="2.46989692s" podCreationTimestamp="2025-11-24 14:38:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:34.462149471 +0000 UTC m=+1151.578789948" watchObservedRunningTime="2025-11-24 14:38:34.46989692 +0000 UTC m=+1151.586537398" Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.139378 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cdd657c7c-xmwlh" Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.258315 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d9f4b5fbd-wzx5t"] Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.259786 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d9f4b5fbd-wzx5t" podUID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerName="neutron-api" containerID="cri-o://f5de7618b4c29ab4b842e9d844eae59251eceff2206035110af1d83066023756" gracePeriod=30 Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.261700 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d9f4b5fbd-wzx5t" podUID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerName="neutron-httpd" containerID="cri-o://786e0d43cffd1b10ee7645509434625f6b501a6e88bc6be1b2f820f28ee2dcd8" gracePeriod=30 Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.465724 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0fbdcd8a-a688-4c39-b968-36c5d3abecbd","Type":"ContainerStarted","Data":"78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7"} Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.478086 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerName="cinder-api-log" containerID="cri-o://ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c" gracePeriod=30 Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.478349 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a0aeb993-7d61-4924-bc31-86c00d8accb5","Type":"ContainerStarted","Data":"1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044"} Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.478830 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.479086 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerName="cinder-api" containerID="cri-o://1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044" gracePeriod=30 Nov 24 14:38:35 crc kubenswrapper[4822]: I1124 14:38:35.509879 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.50986297 podStartE2EDuration="4.50986297s" podCreationTimestamp="2025-11-24 14:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:35.505911298 +0000 UTC m=+1152.622551775" watchObservedRunningTime="2025-11-24 14:38:35.50986297 +0000 UTC m=+1152.626503447" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.509420 4822 generic.go:334] "Generic (PLEG): container finished" podID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerID="786e0d43cffd1b10ee7645509434625f6b501a6e88bc6be1b2f820f28ee2dcd8" exitCode=0 Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.509508 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9f4b5fbd-wzx5t" event={"ID":"e86a7914-dd9a-4b4e-abd7-86d055505255","Type":"ContainerDied","Data":"786e0d43cffd1b10ee7645509434625f6b501a6e88bc6be1b2f820f28ee2dcd8"} Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.515384 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0fbdcd8a-a688-4c39-b968-36c5d3abecbd","Type":"ContainerStarted","Data":"c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92"} Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.527429 4822 generic.go:334] "Generic (PLEG): container finished" podID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerID="ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c" exitCode=143 Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.527752 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a0aeb993-7d61-4924-bc31-86c00d8accb5","Type":"ContainerDied","Data":"ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c"} Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.527867 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.537026 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.562822555 podStartE2EDuration="5.537006104s" podCreationTimestamp="2025-11-24 14:38:31 +0000 UTC" firstStartedPulling="2025-11-24 14:38:33.101361224 +0000 UTC m=+1150.218001701" lastFinishedPulling="2025-11-24 14:38:34.075544773 +0000 UTC m=+1151.192185250" observedRunningTime="2025-11-24 14:38:36.532505506 +0000 UTC m=+1153.649146003" watchObservedRunningTime="2025-11-24 14:38:36.537006104 +0000 UTC m=+1153.653646581" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.554603 4822 generic.go:334] "Generic (PLEG): container finished" podID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerID="13e53dd3c7375c0c69232d41aa2e75146658d67590f39698fddd3dc82d27f7a8" exitCode=0 Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.554669 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerDied","Data":"13e53dd3c7375c0c69232d41aa2e75146658d67590f39698fddd3dc82d27f7a8"} Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.637706 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5cbff998cd-sdjjk" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.666656 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.696008 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-run-httpd\") pod \"1ae8141e-86cb-43bc-9ada-307337b1566f\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.696072 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-log-httpd\") pod \"1ae8141e-86cb-43bc-9ada-307337b1566f\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.696151 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2khl7\" (UniqueName: \"kubernetes.io/projected/1ae8141e-86cb-43bc-9ada-307337b1566f-kube-api-access-2khl7\") pod \"1ae8141e-86cb-43bc-9ada-307337b1566f\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.696173 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-config-data\") pod \"1ae8141e-86cb-43bc-9ada-307337b1566f\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.696308 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-sg-core-conf-yaml\") pod \"1ae8141e-86cb-43bc-9ada-307337b1566f\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.696331 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-combined-ca-bundle\") pod \"1ae8141e-86cb-43bc-9ada-307337b1566f\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.696404 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-scripts\") pod \"1ae8141e-86cb-43bc-9ada-307337b1566f\" (UID: \"1ae8141e-86cb-43bc-9ada-307337b1566f\") " Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.707867 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1ae8141e-86cb-43bc-9ada-307337b1566f" (UID: "1ae8141e-86cb-43bc-9ada-307337b1566f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.708146 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1ae8141e-86cb-43bc-9ada-307337b1566f" (UID: "1ae8141e-86cb-43bc-9ada-307337b1566f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.732403 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-scripts" (OuterVolumeSpecName: "scripts") pod "1ae8141e-86cb-43bc-9ada-307337b1566f" (UID: "1ae8141e-86cb-43bc-9ada-307337b1566f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.732546 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ae8141e-86cb-43bc-9ada-307337b1566f-kube-api-access-2khl7" (OuterVolumeSpecName: "kube-api-access-2khl7") pod "1ae8141e-86cb-43bc-9ada-307337b1566f" (UID: "1ae8141e-86cb-43bc-9ada-307337b1566f"). InnerVolumeSpecName "kube-api-access-2khl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.743168 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c9b79f5f4-m48xx"] Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.743471 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c9b79f5f4-m48xx" podUID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerName="barbican-api-log" containerID="cri-o://e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc" gracePeriod=30 Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.745222 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c9b79f5f4-m48xx" podUID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerName="barbican-api" containerID="cri-o://379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b" gracePeriod=30 Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.755348 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1ae8141e-86cb-43bc-9ada-307337b1566f" (UID: "1ae8141e-86cb-43bc-9ada-307337b1566f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.814386 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.814652 4822 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.814663 4822 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ae8141e-86cb-43bc-9ada-307337b1566f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.814673 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2khl7\" (UniqueName: \"kubernetes.io/projected/1ae8141e-86cb-43bc-9ada-307337b1566f-kube-api-access-2khl7\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.814685 4822 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.885194 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ae8141e-86cb-43bc-9ada-307337b1566f" (UID: "1ae8141e-86cb-43bc-9ada-307337b1566f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.896910 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-config-data" (OuterVolumeSpecName: "config-data") pod "1ae8141e-86cb-43bc-9ada-307337b1566f" (UID: "1ae8141e-86cb-43bc-9ada-307337b1566f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.917990 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:36 crc kubenswrapper[4822]: I1124 14:38:36.918023 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae8141e-86cb-43bc-9ada-307337b1566f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.047648 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.570597 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.570598 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ae8141e-86cb-43bc-9ada-307337b1566f","Type":"ContainerDied","Data":"ba93bc80bc96c1fe1bf2b47001c78988d78b1a1a8aaeac6cedc5cdd15cca7d84"} Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.571277 4822 scope.go:117] "RemoveContainer" containerID="fcffa52afa71bb22a16f640ca7bc1595c95e0c1623ceda38c4d474445aa8eb81" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.572918 4822 generic.go:334] "Generic (PLEG): container finished" podID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerID="e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc" exitCode=143 Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.572975 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9b79f5f4-m48xx" event={"ID":"23b52820-32f7-4734-a60d-c7d0b21206b5","Type":"ContainerDied","Data":"e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc"} Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.614816 4822 scope.go:117] "RemoveContainer" containerID="6fd3b5a8ca9cfaa068d4131029730208157f302f41fbcd2df94c39285f3a9420" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.629400 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.664602 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.683236 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:38:37 crc kubenswrapper[4822]: E1124 14:38:37.683720 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="ceilometer-notification-agent" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.683744 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="ceilometer-notification-agent" Nov 24 14:38:37 crc kubenswrapper[4822]: E1124 14:38:37.683770 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="sg-core" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.683779 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="sg-core" Nov 24 14:38:37 crc kubenswrapper[4822]: E1124 14:38:37.683810 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="proxy-httpd" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.683819 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="proxy-httpd" Nov 24 14:38:37 crc kubenswrapper[4822]: E1124 14:38:37.683842 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="ceilometer-central-agent" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.683851 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="ceilometer-central-agent" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.684086 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="ceilometer-central-agent" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.684121 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="proxy-httpd" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.684137 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="ceilometer-notification-agent" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.684149 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" containerName="sg-core" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.690678 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.692734 4822 scope.go:117] "RemoveContainer" containerID="13e53dd3c7375c0c69232d41aa2e75146658d67590f39698fddd3dc82d27f7a8" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.693746 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.693920 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.728727 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ae8141e-86cb-43bc-9ada-307337b1566f" path="/var/lib/kubelet/pods/1ae8141e-86cb-43bc-9ada-307337b1566f/volumes" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.730338 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.735515 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-run-httpd\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.735549 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-log-httpd\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.735585 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vrz9\" (UniqueName: \"kubernetes.io/projected/be01d337-ad72-4656-82aa-33edea15526b-kube-api-access-7vrz9\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.735621 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-scripts\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.735643 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.735693 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.735730 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-config-data\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.757933 4822 scope.go:117] "RemoveContainer" containerID="f9a32f0cc45408679a748bfb9c5fe147c7e9c92bf6130f47e7e57fe7e0b2eaa7" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.836897 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-config-data\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.836995 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-run-httpd\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.837016 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-log-httpd\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.837049 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vrz9\" (UniqueName: \"kubernetes.io/projected/be01d337-ad72-4656-82aa-33edea15526b-kube-api-access-7vrz9\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.837087 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-scripts\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.837107 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.837155 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.840688 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-log-httpd\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.841100 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-run-httpd\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.843837 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-scripts\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.844725 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.844887 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.845408 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-config-data\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:37 crc kubenswrapper[4822]: I1124 14:38:37.864135 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vrz9\" (UniqueName: \"kubernetes.io/projected/be01d337-ad72-4656-82aa-33edea15526b-kube-api-access-7vrz9\") pod \"ceilometer-0\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " pod="openstack/ceilometer-0" Nov 24 14:38:38 crc kubenswrapper[4822]: I1124 14:38:38.028045 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:38:38 crc kubenswrapper[4822]: I1124 14:38:38.564025 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:38:38 crc kubenswrapper[4822]: I1124 14:38:38.594823 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerStarted","Data":"44eebce83c63d97b358b67832009ae814922e304a696f7c05748f63057cae5b0"} Nov 24 14:38:38 crc kubenswrapper[4822]: I1124 14:38:38.596387 4822 generic.go:334] "Generic (PLEG): container finished" podID="c25c6942-9f38-4db4-834f-2262ee899ed7" containerID="16409edc9499b818433aed5a90f6f114cc12534d16259515b0d3505fd5a7f229" exitCode=0 Nov 24 14:38:38 crc kubenswrapper[4822]: I1124 14:38:38.596494 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-j72lq" event={"ID":"c25c6942-9f38-4db4-834f-2262ee899ed7","Type":"ContainerDied","Data":"16409edc9499b818433aed5a90f6f114cc12534d16259515b0d3505fd5a7f229"} Nov 24 14:38:39 crc kubenswrapper[4822]: I1124 14:38:39.611840 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerStarted","Data":"665a04c41c091fde310ae40cf0689f2e8e9ccabbc82fc659605155c48fda2444"} Nov 24 14:38:39 crc kubenswrapper[4822]: E1124 14:38:39.986668 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="4acc7e6a-472b-468a-b709-183f8b3c2b5b" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.144680 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.292025 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8v2d\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-kube-api-access-t8v2d\") pod \"c25c6942-9f38-4db4-834f-2262ee899ed7\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.293111 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-scripts\") pod \"c25c6942-9f38-4db4-834f-2262ee899ed7\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.293162 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-certs\") pod \"c25c6942-9f38-4db4-834f-2262ee899ed7\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.293178 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-combined-ca-bundle\") pod \"c25c6942-9f38-4db4-834f-2262ee899ed7\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.293301 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-config-data\") pod \"c25c6942-9f38-4db4-834f-2262ee899ed7\" (UID: \"c25c6942-9f38-4db4-834f-2262ee899ed7\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.296757 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-kube-api-access-t8v2d" (OuterVolumeSpecName: "kube-api-access-t8v2d") pod "c25c6942-9f38-4db4-834f-2262ee899ed7" (UID: "c25c6942-9f38-4db4-834f-2262ee899ed7"). InnerVolumeSpecName "kube-api-access-t8v2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.299883 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-scripts" (OuterVolumeSpecName: "scripts") pod "c25c6942-9f38-4db4-834f-2262ee899ed7" (UID: "c25c6942-9f38-4db4-834f-2262ee899ed7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.303899 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-certs" (OuterVolumeSpecName: "certs") pod "c25c6942-9f38-4db4-834f-2262ee899ed7" (UID: "c25c6942-9f38-4db4-834f-2262ee899ed7"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.326054 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-config-data" (OuterVolumeSpecName: "config-data") pod "c25c6942-9f38-4db4-834f-2262ee899ed7" (UID: "c25c6942-9f38-4db4-834f-2262ee899ed7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.336613 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c25c6942-9f38-4db4-834f-2262ee899ed7" (UID: "c25c6942-9f38-4db4-834f-2262ee899ed7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.396298 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8v2d\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-kube-api-access-t8v2d\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.396332 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.396350 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.396361 4822 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/c25c6942-9f38-4db4-834f-2262ee899ed7-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.396373 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c25c6942-9f38-4db4-834f-2262ee899ed7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.401541 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.497045 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data\") pod \"23b52820-32f7-4734-a60d-c7d0b21206b5\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.497198 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b52820-32f7-4734-a60d-c7d0b21206b5-logs\") pod \"23b52820-32f7-4734-a60d-c7d0b21206b5\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.497286 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnhpp\" (UniqueName: \"kubernetes.io/projected/23b52820-32f7-4734-a60d-c7d0b21206b5-kube-api-access-vnhpp\") pod \"23b52820-32f7-4734-a60d-c7d0b21206b5\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.497308 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data-custom\") pod \"23b52820-32f7-4734-a60d-c7d0b21206b5\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.497384 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-combined-ca-bundle\") pod \"23b52820-32f7-4734-a60d-c7d0b21206b5\" (UID: \"23b52820-32f7-4734-a60d-c7d0b21206b5\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.497651 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23b52820-32f7-4734-a60d-c7d0b21206b5-logs" (OuterVolumeSpecName: "logs") pod "23b52820-32f7-4734-a60d-c7d0b21206b5" (UID: "23b52820-32f7-4734-a60d-c7d0b21206b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.497950 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b52820-32f7-4734-a60d-c7d0b21206b5-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.502076 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "23b52820-32f7-4734-a60d-c7d0b21206b5" (UID: "23b52820-32f7-4734-a60d-c7d0b21206b5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.515413 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23b52820-32f7-4734-a60d-c7d0b21206b5-kube-api-access-vnhpp" (OuterVolumeSpecName: "kube-api-access-vnhpp") pod "23b52820-32f7-4734-a60d-c7d0b21206b5" (UID: "23b52820-32f7-4734-a60d-c7d0b21206b5"). InnerVolumeSpecName "kube-api-access-vnhpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.547907 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23b52820-32f7-4734-a60d-c7d0b21206b5" (UID: "23b52820-32f7-4734-a60d-c7d0b21206b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.565162 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data" (OuterVolumeSpecName: "config-data") pod "23b52820-32f7-4734-a60d-c7d0b21206b5" (UID: "23b52820-32f7-4734-a60d-c7d0b21206b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.599302 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnhpp\" (UniqueName: \"kubernetes.io/projected/23b52820-32f7-4734-a60d-c7d0b21206b5-kube-api-access-vnhpp\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.599339 4822 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.599352 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.599375 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b52820-32f7-4734-a60d-c7d0b21206b5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.625582 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerStarted","Data":"b6799e2d2ec8963b470f1896ff5c09f2582ee76336dc78bdb0be4273b1d43f6a"} Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.627492 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-j72lq" event={"ID":"c25c6942-9f38-4db4-834f-2262ee899ed7","Type":"ContainerDied","Data":"311ae1113bb8da0e8915ceea0937843d572bf289ce90e06ad719fa1ca6a4edc7"} Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.627532 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="311ae1113bb8da0e8915ceea0937843d572bf289ce90e06ad719fa1ca6a4edc7" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.627574 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-j72lq" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.631250 4822 generic.go:334] "Generic (PLEG): container finished" podID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerID="f5de7618b4c29ab4b842e9d844eae59251eceff2206035110af1d83066023756" exitCode=0 Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.631301 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9f4b5fbd-wzx5t" event={"ID":"e86a7914-dd9a-4b4e-abd7-86d055505255","Type":"ContainerDied","Data":"f5de7618b4c29ab4b842e9d844eae59251eceff2206035110af1d83066023756"} Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.633575 4822 generic.go:334] "Generic (PLEG): container finished" podID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerID="379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b" exitCode=0 Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.633685 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.633824 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9b79f5f4-m48xx" event={"ID":"23b52820-32f7-4734-a60d-c7d0b21206b5","Type":"ContainerDied","Data":"379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b"} Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.633878 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9b79f5f4-m48xx" event={"ID":"23b52820-32f7-4734-a60d-c7d0b21206b5","Type":"ContainerDied","Data":"9f202d44f968d4db6fb912f3643e1db192ddc3e6c485c55c57e287d08d9efbd0"} Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.633898 4822 scope.go:117] "RemoveContainer" containerID="379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.634133 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9b79f5f4-m48xx" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.671795 4822 scope.go:117] "RemoveContainer" containerID="e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.686397 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c9b79f5f4-m48xx"] Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.697147 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6c9b79f5f4-m48xx"] Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.697331 4822 scope.go:117] "RemoveContainer" containerID="379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b" Nov 24 14:38:40 crc kubenswrapper[4822]: E1124 14:38:40.697743 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b\": container with ID starting with 379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b not found: ID does not exist" containerID="379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.697780 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b"} err="failed to get container status \"379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b\": rpc error: code = NotFound desc = could not find container \"379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b\": container with ID starting with 379bb246776d5aeb968f9dd23ab94611ef3ca9b1ef9f328f707044ef37ea0f1b not found: ID does not exist" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.697805 4822 scope.go:117] "RemoveContainer" containerID="e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc" Nov 24 14:38:40 crc kubenswrapper[4822]: E1124 14:38:40.698052 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc\": container with ID starting with e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc not found: ID does not exist" containerID="e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.698086 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc"} err="failed to get container status \"e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc\": rpc error: code = NotFound desc = could not find container \"e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc\": container with ID starting with e38ba2f063253ea02bc0de309a7be8ff1f38470b59f0cdd3a3b7157af791f1dc not found: ID does not exist" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.771422 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899305 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:38:40 crc kubenswrapper[4822]: E1124 14:38:40.899666 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerName="barbican-api" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899683 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerName="barbican-api" Nov 24 14:38:40 crc kubenswrapper[4822]: E1124 14:38:40.899700 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25c6942-9f38-4db4-834f-2262ee899ed7" containerName="cloudkitty-storageinit" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899706 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25c6942-9f38-4db4-834f-2262ee899ed7" containerName="cloudkitty-storageinit" Nov 24 14:38:40 crc kubenswrapper[4822]: E1124 14:38:40.899730 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerName="neutron-httpd" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899736 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerName="neutron-httpd" Nov 24 14:38:40 crc kubenswrapper[4822]: E1124 14:38:40.899747 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerName="barbican-api-log" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899753 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerName="barbican-api-log" Nov 24 14:38:40 crc kubenswrapper[4822]: E1124 14:38:40.899766 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerName="neutron-api" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899773 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerName="neutron-api" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899935 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerName="barbican-api" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899954 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="23b52820-32f7-4734-a60d-c7d0b21206b5" containerName="barbican-api-log" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899973 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="c25c6942-9f38-4db4-834f-2262ee899ed7" containerName="cloudkitty-storageinit" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899984 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerName="neutron-api" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.899994 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="e86a7914-dd9a-4b4e-abd7-86d055505255" containerName="neutron-httpd" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.900653 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.906665 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.906905 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsnks\" (UniqueName: \"kubernetes.io/projected/e86a7914-dd9a-4b4e-abd7-86d055505255-kube-api-access-tsnks\") pod \"e86a7914-dd9a-4b4e-abd7-86d055505255\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.907015 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-combined-ca-bundle\") pod \"e86a7914-dd9a-4b4e-abd7-86d055505255\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.907043 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-httpd-config\") pod \"e86a7914-dd9a-4b4e-abd7-86d055505255\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.907180 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-config\") pod \"e86a7914-dd9a-4b4e-abd7-86d055505255\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.907262 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-ovndb-tls-certs\") pod \"e86a7914-dd9a-4b4e-abd7-86d055505255\" (UID: \"e86a7914-dd9a-4b4e-abd7-86d055505255\") " Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.909797 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-zhp7h" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.910067 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.910294 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.910470 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.919611 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e86a7914-dd9a-4b4e-abd7-86d055505255-kube-api-access-tsnks" (OuterVolumeSpecName: "kube-api-access-tsnks") pod "e86a7914-dd9a-4b4e-abd7-86d055505255" (UID: "e86a7914-dd9a-4b4e-abd7-86d055505255"). InnerVolumeSpecName "kube-api-access-tsnks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.924533 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e86a7914-dd9a-4b4e-abd7-86d055505255" (UID: "e86a7914-dd9a-4b4e-abd7-86d055505255"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.932722 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.968954 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57fff66767-fnp4z"] Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.969191 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" podUID="8610fa72-47f6-4d01-be06-b1d3486c398b" containerName="dnsmasq-dns" containerID="cri-o://fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8" gracePeriod=10 Nov 24 14:38:40 crc kubenswrapper[4822]: I1124 14:38:40.971387 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.000737 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c56c4ff7-2xkkd"] Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.010213 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.014789 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.015111 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.015259 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.015319 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-scripts\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.015335 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-certs\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.015393 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zrqm\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-kube-api-access-7zrqm\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.015586 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsnks\" (UniqueName: \"kubernetes.io/projected/e86a7914-dd9a-4b4e-abd7-86d055505255-kube-api-access-tsnks\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.015598 4822 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.029717 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c56c4ff7-2xkkd"] Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.107081 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e86a7914-dd9a-4b4e-abd7-86d055505255" (UID: "e86a7914-dd9a-4b4e-abd7-86d055505255"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119516 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119588 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-scripts\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119613 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-certs\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119643 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bdxt\" (UniqueName: \"kubernetes.io/projected/f6d5fe88-9544-489a-8fab-1b0ed7781043-kube-api-access-6bdxt\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119684 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zrqm\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-kube-api-access-7zrqm\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119740 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-sb\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119771 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-dns-svc\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119813 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-config\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119895 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.119964 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-nb\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.120045 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.120138 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.128056 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-scripts\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.131741 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.132220 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-certs\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.144247 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.147326 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e86a7914-dd9a-4b4e-abd7-86d055505255" (UID: "e86a7914-dd9a-4b4e-abd7-86d055505255"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.154084 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.156218 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.159908 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.160295 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.171054 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zrqm\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-kube-api-access-7zrqm\") pod \"cloudkitty-proc-0\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.176719 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.187549 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-config" (OuterVolumeSpecName: "config") pod "e86a7914-dd9a-4b4e-abd7-86d055505255" (UID: "e86a7914-dd9a-4b4e-abd7-86d055505255"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.227645 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-nb\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.227769 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bdxt\" (UniqueName: \"kubernetes.io/projected/f6d5fe88-9544-489a-8fab-1b0ed7781043-kube-api-access-6bdxt\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.227817 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-sb\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.227838 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-dns-svc\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.227863 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-config\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.227965 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.227977 4822 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86a7914-dd9a-4b4e-abd7-86d055505255-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.228832 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-config\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.229660 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-dns-svc\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.232688 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-sb\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.233327 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-nb\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.233790 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.246274 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bdxt\" (UniqueName: \"kubernetes.io/projected/f6d5fe88-9544-489a-8fab-1b0ed7781043-kube-api-access-6bdxt\") pod \"dnsmasq-dns-75c56c4ff7-2xkkd\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.331237 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.331295 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-scripts\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.331317 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.331345 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr5wv\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-kube-api-access-sr5wv\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.331377 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-logs\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.331403 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.331423 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-certs\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.386098 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.433710 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.433764 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-scripts\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.433788 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.434592 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr5wv\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-kube-api-access-sr5wv\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.434638 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-logs\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.434657 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.434678 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-certs\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.436707 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-logs\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.440398 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.441913 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-certs\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.444889 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.447657 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-scripts\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.450066 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.456008 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr5wv\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-kube-api-access-sr5wv\") pod \"cloudkitty-api-0\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.497574 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.581520 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.581582 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.581632 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.582559 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb98f980c90c3e30ffa6cd4e582af8567a0ef6fab6806bdeaf156e64abf982bb"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.582619 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://eb98f980c90c3e30ffa6cd4e582af8567a0ef6fab6806bdeaf156e64abf982bb" gracePeriod=600 Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.619617 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.736620 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23b52820-32f7-4734-a60d-c7d0b21206b5" path="/var/lib/kubelet/pods/23b52820-32f7-4734-a60d-c7d0b21206b5/volumes" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.742105 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smnpf\" (UniqueName: \"kubernetes.io/projected/8610fa72-47f6-4d01-be06-b1d3486c398b-kube-api-access-smnpf\") pod \"8610fa72-47f6-4d01-be06-b1d3486c398b\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.742266 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-config\") pod \"8610fa72-47f6-4d01-be06-b1d3486c398b\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.742385 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-dns-svc\") pod \"8610fa72-47f6-4d01-be06-b1d3486c398b\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.742413 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-nb\") pod \"8610fa72-47f6-4d01-be06-b1d3486c398b\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.742474 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-sb\") pod \"8610fa72-47f6-4d01-be06-b1d3486c398b\" (UID: \"8610fa72-47f6-4d01-be06-b1d3486c398b\") " Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.762037 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9f4b5fbd-wzx5t" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.788454 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8610fa72-47f6-4d01-be06-b1d3486c398b-kube-api-access-smnpf" (OuterVolumeSpecName: "kube-api-access-smnpf") pod "8610fa72-47f6-4d01-be06-b1d3486c398b" (UID: "8610fa72-47f6-4d01-be06-b1d3486c398b"). InnerVolumeSpecName "kube-api-access-smnpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.818780 4822 generic.go:334] "Generic (PLEG): container finished" podID="8610fa72-47f6-4d01-be06-b1d3486c398b" containerID="fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8" exitCode=0 Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.818933 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.844436 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smnpf\" (UniqueName: \"kubernetes.io/projected/8610fa72-47f6-4d01-be06-b1d3486c398b-kube-api-access-smnpf\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.870346 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerStarted","Data":"86c4fcfee95aa28d4015bfda8accace4b928a4f937980afa151961aa0876fb91"} Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.870385 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.870402 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9f4b5fbd-wzx5t" event={"ID":"e86a7914-dd9a-4b4e-abd7-86d055505255","Type":"ContainerDied","Data":"f76b7c7767ce3d8d3322fa4494dc9ed25211873c367261fae919193a8454d8cf"} Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.870420 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" event={"ID":"8610fa72-47f6-4d01-be06-b1d3486c398b","Type":"ContainerDied","Data":"fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8"} Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.870432 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57fff66767-fnp4z" event={"ID":"8610fa72-47f6-4d01-be06-b1d3486c398b","Type":"ContainerDied","Data":"ab3996d3ec43ea51e826579d69c252cdb3a909048d3af274bf5b5c57f1c98752"} Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.870451 4822 scope.go:117] "RemoveContainer" containerID="786e0d43cffd1b10ee7645509434625f6b501a6e88bc6be1b2f820f28ee2dcd8" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.936708 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8610fa72-47f6-4d01-be06-b1d3486c398b" (UID: "8610fa72-47f6-4d01-be06-b1d3486c398b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.947439 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.978377 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-config" (OuterVolumeSpecName: "config") pod "8610fa72-47f6-4d01-be06-b1d3486c398b" (UID: "8610fa72-47f6-4d01-be06-b1d3486c398b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.979855 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8610fa72-47f6-4d01-be06-b1d3486c398b" (UID: "8610fa72-47f6-4d01-be06-b1d3486c398b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.987729 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d9f4b5fbd-wzx5t"] Nov 24 14:38:41 crc kubenswrapper[4822]: I1124 14:38:41.991602 4822 scope.go:117] "RemoveContainer" containerID="f5de7618b4c29ab4b842e9d844eae59251eceff2206035110af1d83066023756" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.010118 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5d9f4b5fbd-wzx5t"] Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.013907 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8610fa72-47f6-4d01-be06-b1d3486c398b" (UID: "8610fa72-47f6-4d01-be06-b1d3486c398b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.044289 4822 scope.go:117] "RemoveContainer" containerID="fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.049325 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.049347 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.049357 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8610fa72-47f6-4d01-be06-b1d3486c398b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.092536 4822 scope.go:117] "RemoveContainer" containerID="09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.129291 4822 scope.go:117] "RemoveContainer" containerID="fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8" Nov 24 14:38:42 crc kubenswrapper[4822]: E1124 14:38:42.129937 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8\": container with ID starting with fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8 not found: ID does not exist" containerID="fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.129984 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8"} err="failed to get container status \"fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8\": rpc error: code = NotFound desc = could not find container \"fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8\": container with ID starting with fc97d7a16feb8d21845b6e865a2e21816bced7384f75deec96904e61d54072e8 not found: ID does not exist" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.130010 4822 scope.go:117] "RemoveContainer" containerID="09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2" Nov 24 14:38:42 crc kubenswrapper[4822]: E1124 14:38:42.130670 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2\": container with ID starting with 09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2 not found: ID does not exist" containerID="09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.130703 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2"} err="failed to get container status \"09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2\": rpc error: code = NotFound desc = could not find container \"09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2\": container with ID starting with 09063f1f7d93927ca3a8abfe34dc3ad082821722093fa883fa0741be775f53c2 not found: ID does not exist" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.164702 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57fff66767-fnp4z"] Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.193255 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57fff66767-fnp4z"] Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.251191 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.381062 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c56c4ff7-2xkkd"] Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.456046 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.492981 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.872052 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" event={"ID":"f6d5fe88-9544-489a-8fab-1b0ed7781043","Type":"ContainerStarted","Data":"06725a3b562a7d0e390d3800840fb5829929c539bd0e1e05e34ca42a8766b748"} Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.872306 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" event={"ID":"f6d5fe88-9544-489a-8fab-1b0ed7781043","Type":"ContainerStarted","Data":"34b3562521fac0f0d263b7b2b9937419f783e6bdd0d6154421b470ffe659f8bd"} Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.875754 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"cfb584d1-8a63-4397-b137-65ac59e8808e","Type":"ContainerStarted","Data":"519e0c7ab52a9e6b186e34f4edd7cd2c1dd528aaf29c4367a61e62ebda81862b"} Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.878002 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerStarted","Data":"d9a2ab24e902dcae258c6c3344acdb65b1fd26e52caba9b4a100dd420871cb26"} Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.878326 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.900857 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"4a3cf934-d11b-48d9-80cb-bdc6efcdf775","Type":"ContainerStarted","Data":"ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138"} Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.900891 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"4a3cf934-d11b-48d9-80cb-bdc6efcdf775","Type":"ContainerStarted","Data":"709da3a4b47960c3da7c2400e294200f290d966422e1c39605b67ff2eec36aed"} Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.904315 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="eb98f980c90c3e30ffa6cd4e582af8567a0ef6fab6806bdeaf156e64abf982bb" exitCode=0 Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.904560 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerName="cinder-scheduler" containerID="cri-o://78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7" gracePeriod=30 Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.904878 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"eb98f980c90c3e30ffa6cd4e582af8567a0ef6fab6806bdeaf156e64abf982bb"} Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.904923 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"a9b7bf9be06a9bc314261db2c2d98effa00916294896221883633ba00a47f5d1"} Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.904949 4822 scope.go:117] "RemoveContainer" containerID="cb8615da8ae4badf56e6f90df575844458ac64f06801fa8ab7085476265bc5df" Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.905039 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerName="probe" containerID="cri-o://c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92" gracePeriod=30 Nov 24 14:38:42 crc kubenswrapper[4822]: I1124 14:38:42.930703 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.267358447 podStartE2EDuration="5.930674367s" podCreationTimestamp="2025-11-24 14:38:37 +0000 UTC" firstStartedPulling="2025-11-24 14:38:38.568880421 +0000 UTC m=+1155.685520898" lastFinishedPulling="2025-11-24 14:38:42.232196341 +0000 UTC m=+1159.348836818" observedRunningTime="2025-11-24 14:38:42.911602369 +0000 UTC m=+1160.028242876" watchObservedRunningTime="2025-11-24 14:38:42.930674367 +0000 UTC m=+1160.047314854" Nov 24 14:38:43 crc kubenswrapper[4822]: I1124 14:38:43.652435 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:38:43 crc kubenswrapper[4822]: I1124 14:38:43.729366 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8610fa72-47f6-4d01-be06-b1d3486c398b" path="/var/lib/kubelet/pods/8610fa72-47f6-4d01-be06-b1d3486c398b/volumes" Nov 24 14:38:43 crc kubenswrapper[4822]: I1124 14:38:43.743322 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e86a7914-dd9a-4b4e-abd7-86d055505255" path="/var/lib/kubelet/pods/e86a7914-dd9a-4b4e-abd7-86d055505255/volumes" Nov 24 14:38:43 crc kubenswrapper[4822]: I1124 14:38:43.928864 4822 generic.go:334] "Generic (PLEG): container finished" podID="f6d5fe88-9544-489a-8fab-1b0ed7781043" containerID="06725a3b562a7d0e390d3800840fb5829929c539bd0e1e05e34ca42a8766b748" exitCode=0 Nov 24 14:38:43 crc kubenswrapper[4822]: I1124 14:38:43.929180 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" event={"ID":"f6d5fe88-9544-489a-8fab-1b0ed7781043","Type":"ContainerDied","Data":"06725a3b562a7d0e390d3800840fb5829929c539bd0e1e05e34ca42a8766b748"} Nov 24 14:38:43 crc kubenswrapper[4822]: I1124 14:38:43.942255 4822 generic.go:334] "Generic (PLEG): container finished" podID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerID="c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92" exitCode=0 Nov 24 14:38:43 crc kubenswrapper[4822]: I1124 14:38:43.943092 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0fbdcd8a-a688-4c39-b968-36c5d3abecbd","Type":"ContainerDied","Data":"c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92"} Nov 24 14:38:43 crc kubenswrapper[4822]: I1124 14:38:43.983509 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=2.287715784 podStartE2EDuration="3.983490824s" podCreationTimestamp="2025-11-24 14:38:40 +0000 UTC" firstStartedPulling="2025-11-24 14:38:41.936859792 +0000 UTC m=+1159.053500269" lastFinishedPulling="2025-11-24 14:38:43.632634832 +0000 UTC m=+1160.749275309" observedRunningTime="2025-11-24 14:38:43.971844764 +0000 UTC m=+1161.088485261" watchObservedRunningTime="2025-11-24 14:38:43.983490824 +0000 UTC m=+1161.100131301" Nov 24 14:38:43 crc kubenswrapper[4822]: I1124 14:38:43.990249 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:38:44 crc kubenswrapper[4822]: I1124 14:38:44.952223 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"4a3cf934-d11b-48d9-80cb-bdc6efcdf775","Type":"ContainerStarted","Data":"ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e"} Nov 24 14:38:44 crc kubenswrapper[4822]: I1124 14:38:44.952334 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerName="cloudkitty-api-log" containerID="cri-o://ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138" gracePeriod=30 Nov 24 14:38:44 crc kubenswrapper[4822]: I1124 14:38:44.953830 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Nov 24 14:38:44 crc kubenswrapper[4822]: I1124 14:38:44.952382 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerName="cloudkitty-api" containerID="cri-o://ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e" gracePeriod=30 Nov 24 14:38:44 crc kubenswrapper[4822]: I1124 14:38:44.956126 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" event={"ID":"f6d5fe88-9544-489a-8fab-1b0ed7781043","Type":"ContainerStarted","Data":"33deb1cb2c718cba129e94fa4cdeebe41fe8b60a7058899d10d63d556af393f3"} Nov 24 14:38:44 crc kubenswrapper[4822]: I1124 14:38:44.956301 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:44 crc kubenswrapper[4822]: I1124 14:38:44.958861 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"cfb584d1-8a63-4397-b137-65ac59e8808e","Type":"ContainerStarted","Data":"98b3025eb88fe0e6ca8ce2e7f0b15fc10da312d8940e44f3b4ba5e3bbaf522ef"} Nov 24 14:38:44 crc kubenswrapper[4822]: I1124 14:38:44.958975 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="cfb584d1-8a63-4397-b137-65ac59e8808e" containerName="cloudkitty-proc" containerID="cri-o://98b3025eb88fe0e6ca8ce2e7f0b15fc10da312d8940e44f3b4ba5e3bbaf522ef" gracePeriod=30 Nov 24 14:38:44 crc kubenswrapper[4822]: I1124 14:38:44.980088 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=3.980069515 podStartE2EDuration="3.980069515s" podCreationTimestamp="2025-11-24 14:38:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:44.974383419 +0000 UTC m=+1162.091023896" watchObservedRunningTime="2025-11-24 14:38:44.980069515 +0000 UTC m=+1162.096710002" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.006558 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" podStartSLOduration=5.006534172 podStartE2EDuration="5.006534172s" podCreationTimestamp="2025-11-24 14:38:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:44.994303064 +0000 UTC m=+1162.110943541" watchObservedRunningTime="2025-11-24 14:38:45.006534172 +0000 UTC m=+1162.123174659" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.009883 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:38:45 crc kubenswrapper[4822]: E1124 14:38:45.010110 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:38:45 crc kubenswrapper[4822]: E1124 14:38:45.010129 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:38:45 crc kubenswrapper[4822]: E1124 14:38:45.010175 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:40:47.010160605 +0000 UTC m=+1284.126801082 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.034307 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.761413 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.929080 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-certs\") pod \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.929170 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-logs\") pod \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.929243 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data\") pod \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.929280 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data-custom\") pod \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.929327 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-combined-ca-bundle\") pod \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.929402 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-scripts\") pod \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.929456 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr5wv\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-kube-api-access-sr5wv\") pod \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\" (UID: \"4a3cf934-d11b-48d9-80cb-bdc6efcdf775\") " Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.929829 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-logs" (OuterVolumeSpecName: "logs") pod "4a3cf934-d11b-48d9-80cb-bdc6efcdf775" (UID: "4a3cf934-d11b-48d9-80cb-bdc6efcdf775"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.938930 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-kube-api-access-sr5wv" (OuterVolumeSpecName: "kube-api-access-sr5wv") pod "4a3cf934-d11b-48d9-80cb-bdc6efcdf775" (UID: "4a3cf934-d11b-48d9-80cb-bdc6efcdf775"). InnerVolumeSpecName "kube-api-access-sr5wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.948420 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4a3cf934-d11b-48d9-80cb-bdc6efcdf775" (UID: "4a3cf934-d11b-48d9-80cb-bdc6efcdf775"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.948558 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-scripts" (OuterVolumeSpecName: "scripts") pod "4a3cf934-d11b-48d9-80cb-bdc6efcdf775" (UID: "4a3cf934-d11b-48d9-80cb-bdc6efcdf775"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.957401 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-certs" (OuterVolumeSpecName: "certs") pod "4a3cf934-d11b-48d9-80cb-bdc6efcdf775" (UID: "4a3cf934-d11b-48d9-80cb-bdc6efcdf775"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.980389 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a3cf934-d11b-48d9-80cb-bdc6efcdf775" (UID: "4a3cf934-d11b-48d9-80cb-bdc6efcdf775"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.987816 4822 generic.go:334] "Generic (PLEG): container finished" podID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerID="ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e" exitCode=0 Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.987929 4822 generic.go:334] "Generic (PLEG): container finished" podID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerID="ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138" exitCode=143 Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.988859 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.989418 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"4a3cf934-d11b-48d9-80cb-bdc6efcdf775","Type":"ContainerDied","Data":"ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e"} Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.989569 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"4a3cf934-d11b-48d9-80cb-bdc6efcdf775","Type":"ContainerDied","Data":"ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138"} Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.989628 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"4a3cf934-d11b-48d9-80cb-bdc6efcdf775","Type":"ContainerDied","Data":"709da3a4b47960c3da7c2400e294200f290d966422e1c39605b67ff2eec36aed"} Nov 24 14:38:45 crc kubenswrapper[4822]: I1124 14:38:45.989699 4822 scope.go:117] "RemoveContainer" containerID="ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.001756 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data" (OuterVolumeSpecName: "config-data") pod "4a3cf934-d11b-48d9-80cb-bdc6efcdf775" (UID: "4a3cf934-d11b-48d9-80cb-bdc6efcdf775"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.032769 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.032799 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.032810 4822 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.032822 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.032831 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.032839 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr5wv\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-kube-api-access-sr5wv\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.032847 4822 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4a3cf934-d11b-48d9-80cb-bdc6efcdf775-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.077704 4822 scope.go:117] "RemoveContainer" containerID="ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.118460 4822 scope.go:117] "RemoveContainer" containerID="ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e" Nov 24 14:38:46 crc kubenswrapper[4822]: E1124 14:38:46.126146 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e\": container with ID starting with ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e not found: ID does not exist" containerID="ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.126187 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e"} err="failed to get container status \"ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e\": rpc error: code = NotFound desc = could not find container \"ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e\": container with ID starting with ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e not found: ID does not exist" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.126226 4822 scope.go:117] "RemoveContainer" containerID="ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138" Nov 24 14:38:46 crc kubenswrapper[4822]: E1124 14:38:46.130286 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138\": container with ID starting with ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138 not found: ID does not exist" containerID="ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.130323 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138"} err="failed to get container status \"ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138\": rpc error: code = NotFound desc = could not find container \"ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138\": container with ID starting with ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138 not found: ID does not exist" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.130344 4822 scope.go:117] "RemoveContainer" containerID="ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.132716 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e"} err="failed to get container status \"ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e\": rpc error: code = NotFound desc = could not find container \"ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e\": container with ID starting with ba4016c18b59cf56874494b0415ffb57d9df6ff2bed324a7712d2bf73bc7621e not found: ID does not exist" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.135462 4822 scope.go:117] "RemoveContainer" containerID="ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.144578 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138"} err="failed to get container status \"ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138\": rpc error: code = NotFound desc = could not find container \"ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138\": container with ID starting with ffdcb44e1ac80d4a63ea689a58f4b0607ebe5dab0d998596be517d1b7bbb8138 not found: ID does not exist" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.329300 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.342271 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.348572 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:38:46 crc kubenswrapper[4822]: E1124 14:38:46.349059 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8610fa72-47f6-4d01-be06-b1d3486c398b" containerName="init" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.349072 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8610fa72-47f6-4d01-be06-b1d3486c398b" containerName="init" Nov 24 14:38:46 crc kubenswrapper[4822]: E1124 14:38:46.349088 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8610fa72-47f6-4d01-be06-b1d3486c398b" containerName="dnsmasq-dns" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.349094 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8610fa72-47f6-4d01-be06-b1d3486c398b" containerName="dnsmasq-dns" Nov 24 14:38:46 crc kubenswrapper[4822]: E1124 14:38:46.349107 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerName="cloudkitty-api-log" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.349114 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerName="cloudkitty-api-log" Nov 24 14:38:46 crc kubenswrapper[4822]: E1124 14:38:46.349142 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerName="cloudkitty-api" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.349150 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerName="cloudkitty-api" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.349373 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerName="cloudkitty-api-log" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.349404 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="8610fa72-47f6-4d01-be06-b1d3486c398b" containerName="dnsmasq-dns" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.349420 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" containerName="cloudkitty-api" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.350541 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.358193 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.358453 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.358603 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.360426 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.440335 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-logs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.440745 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxx8s\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-kube-api-access-hxx8s\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.440790 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.440836 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.440862 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.440890 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.440918 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-certs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.440936 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.440983 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-scripts\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.544058 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.544139 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.544177 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.544234 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-certs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.544262 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.544321 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-scripts\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.544540 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-logs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.544614 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxx8s\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-kube-api-access-hxx8s\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.562616 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-logs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.565323 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.579946 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.582597 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.581700 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-scripts\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.589455 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-certs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.589943 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.591838 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.597347 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.604192 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxx8s\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-kube-api-access-hxx8s\") pod \"cloudkitty-api-0\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " pod="openstack/cloudkitty-api-0" Nov 24 14:38:46 crc kubenswrapper[4822]: I1124 14:38:46.676888 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.104911 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.553587 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.712024 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-etc-machine-id\") pod \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.712621 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-combined-ca-bundle\") pod \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.712709 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2vft\" (UniqueName: \"kubernetes.io/projected/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-kube-api-access-j2vft\") pod \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.712776 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data\") pod \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.712853 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data-custom\") pod \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.713070 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-scripts\") pod \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\" (UID: \"0fbdcd8a-a688-4c39-b968-36c5d3abecbd\") " Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.720151 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-kube-api-access-j2vft" (OuterVolumeSpecName: "kube-api-access-j2vft") pod "0fbdcd8a-a688-4c39-b968-36c5d3abecbd" (UID: "0fbdcd8a-a688-4c39-b968-36c5d3abecbd"). InnerVolumeSpecName "kube-api-access-j2vft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.720295 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0fbdcd8a-a688-4c39-b968-36c5d3abecbd" (UID: "0fbdcd8a-a688-4c39-b968-36c5d3abecbd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.720618 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a3cf934-d11b-48d9-80cb-bdc6efcdf775" path="/var/lib/kubelet/pods/4a3cf934-d11b-48d9-80cb-bdc6efcdf775/volumes" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.725025 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0fbdcd8a-a688-4c39-b968-36c5d3abecbd" (UID: "0fbdcd8a-a688-4c39-b968-36c5d3abecbd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.729848 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-scripts" (OuterVolumeSpecName: "scripts") pod "0fbdcd8a-a688-4c39-b968-36c5d3abecbd" (UID: "0fbdcd8a-a688-4c39-b968-36c5d3abecbd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.801932 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fbdcd8a-a688-4c39-b968-36c5d3abecbd" (UID: "0fbdcd8a-a688-4c39-b968-36c5d3abecbd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.819509 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.819535 4822 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.819546 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.819555 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2vft\" (UniqueName: \"kubernetes.io/projected/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-kube-api-access-j2vft\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.819566 4822 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.857033 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data" (OuterVolumeSpecName: "config-data") pod "0fbdcd8a-a688-4c39-b968-36c5d3abecbd" (UID: "0fbdcd8a-a688-4c39-b968-36c5d3abecbd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:38:47 crc kubenswrapper[4822]: I1124 14:38:47.920966 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdcd8a-a688-4c39-b968-36c5d3abecbd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.009045 4822 generic.go:334] "Generic (PLEG): container finished" podID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerID="78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7" exitCode=0 Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.009132 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0fbdcd8a-a688-4c39-b968-36c5d3abecbd","Type":"ContainerDied","Data":"78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7"} Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.009166 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0fbdcd8a-a688-4c39-b968-36c5d3abecbd","Type":"ContainerDied","Data":"4ea03ac80448c094bcb137c517c41f8967174f22ebb994fc6266ab43215094a4"} Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.009186 4822 scope.go:117] "RemoveContainer" containerID="c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.009333 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.014072 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"51daf6fb-63aa-4ab7-8e3b-69e861e67e43","Type":"ContainerStarted","Data":"a8fe55b14cbb0e20e3997f9943574c2356792e3425cd3798ec3cc8bfa3a82050"} Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.014294 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.014374 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"51daf6fb-63aa-4ab7-8e3b-69e861e67e43","Type":"ContainerStarted","Data":"0ed5ee75ecb4ed8f0fbfe3dff427d71463d18ecc6ae018bce84cf9fcbe2de39b"} Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.014433 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"51daf6fb-63aa-4ab7-8e3b-69e861e67e43","Type":"ContainerStarted","Data":"d8dc33f4986b6160e6dccc5041abadad819e63c306e0c2fd51e4a684fcbb37e2"} Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.042798 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.042776453 podStartE2EDuration="2.042776453s" podCreationTimestamp="2025-11-24 14:38:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:48.030754957 +0000 UTC m=+1165.147395454" watchObservedRunningTime="2025-11-24 14:38:48.042776453 +0000 UTC m=+1165.159416930" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.055011 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.062549 4822 scope.go:117] "RemoveContainer" containerID="78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.066356 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.081052 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:38:48 crc kubenswrapper[4822]: E1124 14:38:48.081527 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerName="cinder-scheduler" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.081544 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerName="cinder-scheduler" Nov 24 14:38:48 crc kubenswrapper[4822]: E1124 14:38:48.081554 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerName="probe" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.081561 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerName="probe" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.081778 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerName="cinder-scheduler" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.081801 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" containerName="probe" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.082882 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.085127 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.087884 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.094367 4822 scope.go:117] "RemoveContainer" containerID="c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92" Nov 24 14:38:48 crc kubenswrapper[4822]: E1124 14:38:48.098313 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92\": container with ID starting with c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92 not found: ID does not exist" containerID="c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.098352 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92"} err="failed to get container status \"c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92\": rpc error: code = NotFound desc = could not find container \"c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92\": container with ID starting with c6948ed89e6e9d348e1c2f22f3f41a3fe0e757303c52e9ad784cfb27be719d92 not found: ID does not exist" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.098375 4822 scope.go:117] "RemoveContainer" containerID="78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7" Nov 24 14:38:48 crc kubenswrapper[4822]: E1124 14:38:48.100250 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7\": container with ID starting with 78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7 not found: ID does not exist" containerID="78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.100275 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7"} err="failed to get container status \"78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7\": rpc error: code = NotFound desc = could not find container \"78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7\": container with ID starting with 78b582ffc7eeaf42919114f3fb617db499eec9727363ed3174b956370f9407a7 not found: ID does not exist" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.226530 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96b0c116-ea8d-4c76-b95d-0b95131da043-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.227689 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9llb6\" (UniqueName: \"kubernetes.io/projected/96b0c116-ea8d-4c76-b95d-0b95131da043-kube-api-access-9llb6\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.227780 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.227868 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.228171 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-config-data\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.228394 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-scripts\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.330401 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.330493 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-config-data\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.330645 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-scripts\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.330754 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96b0c116-ea8d-4c76-b95d-0b95131da043-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.330871 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9llb6\" (UniqueName: \"kubernetes.io/projected/96b0c116-ea8d-4c76-b95d-0b95131da043-kube-api-access-9llb6\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.330918 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.331013 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96b0c116-ea8d-4c76-b95d-0b95131da043-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.336989 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-scripts\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.337064 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.340382 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-config-data\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.345334 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96b0c116-ea8d-4c76-b95d-0b95131da043-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.358348 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9llb6\" (UniqueName: \"kubernetes.io/projected/96b0c116-ea8d-4c76-b95d-0b95131da043-kube-api-access-9llb6\") pod \"cinder-scheduler-0\" (UID: \"96b0c116-ea8d-4c76-b95d-0b95131da043\") " pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.413391 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:38:48 crc kubenswrapper[4822]: I1124 14:38:48.900657 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:38:49 crc kubenswrapper[4822]: I1124 14:38:49.032946 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"96b0c116-ea8d-4c76-b95d-0b95131da043","Type":"ContainerStarted","Data":"006a2097d12dbc37dbb79f1599426dcd3cf1d0ab14c08e9ec4889ef5c0c2cb38"} Nov 24 14:38:49 crc kubenswrapper[4822]: I1124 14:38:49.725767 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fbdcd8a-a688-4c39-b968-36c5d3abecbd" path="/var/lib/kubelet/pods/0fbdcd8a-a688-4c39-b968-36c5d3abecbd/volumes" Nov 24 14:38:50 crc kubenswrapper[4822]: I1124 14:38:50.046767 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"96b0c116-ea8d-4c76-b95d-0b95131da043","Type":"ContainerStarted","Data":"20600fbc1f9f48dac0c490e03247e75ac26eb9c36de114fc785faf07ca71eba9"} Nov 24 14:38:51 crc kubenswrapper[4822]: I1124 14:38:51.058270 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"96b0c116-ea8d-4c76-b95d-0b95131da043","Type":"ContainerStarted","Data":"66ee6f007d72a396d2dfb0520372cb0c8cd86f5449a54c5620aba9a2462ab8d7"} Nov 24 14:38:51 crc kubenswrapper[4822]: I1124 14:38:51.412493 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:38:51 crc kubenswrapper[4822]: I1124 14:38:51.441001 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.440964924 podStartE2EDuration="3.440964924s" podCreationTimestamp="2025-11-24 14:38:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:51.078171516 +0000 UTC m=+1168.194812033" watchObservedRunningTime="2025-11-24 14:38:51.440964924 +0000 UTC m=+1168.557605441" Nov 24 14:38:51 crc kubenswrapper[4822]: I1124 14:38:51.468492 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d649d8c65-ptdtg"] Nov 24 14:38:51 crc kubenswrapper[4822]: I1124 14:38:51.468738 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" podUID="fd67711e-c7f9-4cc8-85e8-1886545615f2" containerName="dnsmasq-dns" containerID="cri-o://610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244" gracePeriod=10 Nov 24 14:38:51 crc kubenswrapper[4822]: I1124 14:38:51.527757 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:51 crc kubenswrapper[4822]: I1124 14:38:51.594476 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d6f884c6d-xlmbh" Nov 24 14:38:51 crc kubenswrapper[4822]: I1124 14:38:51.912672 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7f958c4db4-6hhpg" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.041109 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.081403 4822 generic.go:334] "Generic (PLEG): container finished" podID="fd67711e-c7f9-4cc8-85e8-1886545615f2" containerID="610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244" exitCode=0 Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.081476 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" event={"ID":"fd67711e-c7f9-4cc8-85e8-1886545615f2","Type":"ContainerDied","Data":"610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244"} Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.081520 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" event={"ID":"fd67711e-c7f9-4cc8-85e8-1886545615f2","Type":"ContainerDied","Data":"ad397f36e0e8906132cbd799780465c54f6bb81c60591cc8b30fc832ae81d551"} Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.081541 4822 scope.go:117] "RemoveContainer" containerID="610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.081655 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d649d8c65-ptdtg" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.108938 4822 scope.go:117] "RemoveContainer" containerID="0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.135607 4822 scope.go:117] "RemoveContainer" containerID="610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244" Nov 24 14:38:52 crc kubenswrapper[4822]: E1124 14:38:52.140264 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244\": container with ID starting with 610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244 not found: ID does not exist" containerID="610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.140305 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244"} err="failed to get container status \"610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244\": rpc error: code = NotFound desc = could not find container \"610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244\": container with ID starting with 610f41bf9d3cfdbc8fab5767a3b61b03aa89ae0a0377322999a561c3731cd244 not found: ID does not exist" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.140327 4822 scope.go:117] "RemoveContainer" containerID="0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24" Nov 24 14:38:52 crc kubenswrapper[4822]: E1124 14:38:52.141632 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24\": container with ID starting with 0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24 not found: ID does not exist" containerID="0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.141683 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24"} err="failed to get container status \"0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24\": rpc error: code = NotFound desc = could not find container \"0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24\": container with ID starting with 0de653800ed1c35f6e95f3973c950f506e8680ad6bdca86b1adac83310180d24 not found: ID does not exist" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.231030 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljqsg\" (UniqueName: \"kubernetes.io/projected/fd67711e-c7f9-4cc8-85e8-1886545615f2-kube-api-access-ljqsg\") pod \"fd67711e-c7f9-4cc8-85e8-1886545615f2\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.231169 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-config\") pod \"fd67711e-c7f9-4cc8-85e8-1886545615f2\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.231251 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-dns-svc\") pod \"fd67711e-c7f9-4cc8-85e8-1886545615f2\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.231304 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-sb\") pod \"fd67711e-c7f9-4cc8-85e8-1886545615f2\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.231334 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-nb\") pod \"fd67711e-c7f9-4cc8-85e8-1886545615f2\" (UID: \"fd67711e-c7f9-4cc8-85e8-1886545615f2\") " Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.239376 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd67711e-c7f9-4cc8-85e8-1886545615f2-kube-api-access-ljqsg" (OuterVolumeSpecName: "kube-api-access-ljqsg") pod "fd67711e-c7f9-4cc8-85e8-1886545615f2" (UID: "fd67711e-c7f9-4cc8-85e8-1886545615f2"). InnerVolumeSpecName "kube-api-access-ljqsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.287867 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fd67711e-c7f9-4cc8-85e8-1886545615f2" (UID: "fd67711e-c7f9-4cc8-85e8-1886545615f2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.294149 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fd67711e-c7f9-4cc8-85e8-1886545615f2" (UID: "fd67711e-c7f9-4cc8-85e8-1886545615f2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.295768 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-config" (OuterVolumeSpecName: "config") pod "fd67711e-c7f9-4cc8-85e8-1886545615f2" (UID: "fd67711e-c7f9-4cc8-85e8-1886545615f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.314040 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fd67711e-c7f9-4cc8-85e8-1886545615f2" (UID: "fd67711e-c7f9-4cc8-85e8-1886545615f2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.334263 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljqsg\" (UniqueName: \"kubernetes.io/projected/fd67711e-c7f9-4cc8-85e8-1886545615f2-kube-api-access-ljqsg\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.334295 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.334306 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.334324 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.334348 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd67711e-c7f9-4cc8-85e8-1886545615f2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.431616 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d649d8c65-ptdtg"] Nov 24 14:38:52 crc kubenswrapper[4822]: I1124 14:38:52.441920 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d649d8c65-ptdtg"] Nov 24 14:38:53 crc kubenswrapper[4822]: I1124 14:38:53.413870 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 14:38:53 crc kubenswrapper[4822]: I1124 14:38:53.715112 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd67711e-c7f9-4cc8-85e8-1886545615f2" path="/var/lib/kubelet/pods/fd67711e-c7f9-4cc8-85e8-1886545615f2/volumes" Nov 24 14:38:54 crc kubenswrapper[4822]: I1124 14:38:54.960484 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 14:38:54 crc kubenswrapper[4822]: E1124 14:38:54.961694 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd67711e-c7f9-4cc8-85e8-1886545615f2" containerName="dnsmasq-dns" Nov 24 14:38:54 crc kubenswrapper[4822]: I1124 14:38:54.961720 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd67711e-c7f9-4cc8-85e8-1886545615f2" containerName="dnsmasq-dns" Nov 24 14:38:54 crc kubenswrapper[4822]: E1124 14:38:54.961777 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd67711e-c7f9-4cc8-85e8-1886545615f2" containerName="init" Nov 24 14:38:54 crc kubenswrapper[4822]: I1124 14:38:54.961788 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd67711e-c7f9-4cc8-85e8-1886545615f2" containerName="init" Nov 24 14:38:54 crc kubenswrapper[4822]: I1124 14:38:54.962136 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd67711e-c7f9-4cc8-85e8-1886545615f2" containerName="dnsmasq-dns" Nov 24 14:38:54 crc kubenswrapper[4822]: I1124 14:38:54.963356 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:38:54 crc kubenswrapper[4822]: I1124 14:38:54.967366 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 14:38:54 crc kubenswrapper[4822]: I1124 14:38:54.967671 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 14:38:54 crc kubenswrapper[4822]: I1124 14:38:54.971341 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-db8hf" Nov 24 14:38:54 crc kubenswrapper[4822]: I1124 14:38:54.974861 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.116053 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/15618ea1-3097-4bd0-aab1-8feedfc75839-openstack-config-secret\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.116100 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/15618ea1-3097-4bd0-aab1-8feedfc75839-openstack-config\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.116519 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcd2c\" (UniqueName: \"kubernetes.io/projected/15618ea1-3097-4bd0-aab1-8feedfc75839-kube-api-access-mcd2c\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.116601 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15618ea1-3097-4bd0-aab1-8feedfc75839-combined-ca-bundle\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.218725 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15618ea1-3097-4bd0-aab1-8feedfc75839-combined-ca-bundle\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.218875 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/15618ea1-3097-4bd0-aab1-8feedfc75839-openstack-config-secret\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.218899 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/15618ea1-3097-4bd0-aab1-8feedfc75839-openstack-config\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.219079 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcd2c\" (UniqueName: \"kubernetes.io/projected/15618ea1-3097-4bd0-aab1-8feedfc75839-kube-api-access-mcd2c\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.219770 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/15618ea1-3097-4bd0-aab1-8feedfc75839-openstack-config\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.226935 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/15618ea1-3097-4bd0-aab1-8feedfc75839-openstack-config-secret\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.230498 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15618ea1-3097-4bd0-aab1-8feedfc75839-combined-ca-bundle\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.239118 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcd2c\" (UniqueName: \"kubernetes.io/projected/15618ea1-3097-4bd0-aab1-8feedfc75839-kube-api-access-mcd2c\") pod \"openstackclient\" (UID: \"15618ea1-3097-4bd0-aab1-8feedfc75839\") " pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.292344 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:38:55 crc kubenswrapper[4822]: W1124 14:38:55.799185 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15618ea1_3097_4bd0_aab1_8feedfc75839.slice/crio-69fac52660eb48c128f1e2383fe391c2c25f0c2719646e540a56fe617d40c4fe WatchSource:0}: Error finding container 69fac52660eb48c128f1e2383fe391c2c25f0c2719646e540a56fe617d40c4fe: Status 404 returned error can't find the container with id 69fac52660eb48c128f1e2383fe391c2c25f0c2719646e540a56fe617d40c4fe Nov 24 14:38:55 crc kubenswrapper[4822]: I1124 14:38:55.801893 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 14:38:56 crc kubenswrapper[4822]: I1124 14:38:56.126127 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"15618ea1-3097-4bd0-aab1-8feedfc75839","Type":"ContainerStarted","Data":"69fac52660eb48c128f1e2383fe391c2c25f0c2719646e540a56fe617d40c4fe"} Nov 24 14:38:58 crc kubenswrapper[4822]: I1124 14:38:58.720116 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.299777 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6b8fb4f557-bkwxl"] Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.301383 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.305457 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.305696 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.305900 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.376920 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6b8fb4f557-bkwxl"] Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.416833 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-internal-tls-certs\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.416911 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-combined-ca-bundle\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.416966 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-log-httpd\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.416985 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-config-data\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.417038 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-run-httpd\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.417126 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.417144 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr2xc\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-kube-api-access-mr2xc\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.417172 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-public-tls-certs\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.518955 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-combined-ca-bundle\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.519040 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-log-httpd\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.519061 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-config-data\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.519089 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-run-httpd\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.519164 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.519182 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr2xc\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-kube-api-access-mr2xc\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.519223 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-public-tls-certs\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.519262 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-internal-tls-certs\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: E1124 14:38:59.523405 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:38:59 crc kubenswrapper[4822]: E1124 14:38:59.523426 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:38:59 crc kubenswrapper[4822]: E1124 14:38:59.523474 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:39:00.023458309 +0000 UTC m=+1177.140098786 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.524127 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-run-httpd\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.534647 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-internal-tls-certs\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.534907 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-config-data\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.535052 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-log-httpd\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.536268 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-combined-ca-bundle\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.536628 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-public-tls-certs\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:38:59 crc kubenswrapper[4822]: I1124 14:38:59.548879 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr2xc\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-kube-api-access-mr2xc\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:39:00 crc kubenswrapper[4822]: I1124 14:39:00.029941 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:39:00 crc kubenswrapper[4822]: E1124 14:39:00.030231 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:39:00 crc kubenswrapper[4822]: E1124 14:39:00.030255 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:39:00 crc kubenswrapper[4822]: E1124 14:39:00.030312 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:39:01.03029576 +0000 UTC m=+1178.146936237 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:39:00 crc kubenswrapper[4822]: I1124 14:39:00.491126 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:00 crc kubenswrapper[4822]: I1124 14:39:00.492244 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="ceilometer-central-agent" containerID="cri-o://665a04c41c091fde310ae40cf0689f2e8e9ccabbc82fc659605155c48fda2444" gracePeriod=30 Nov 24 14:39:00 crc kubenswrapper[4822]: I1124 14:39:00.492891 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="proxy-httpd" containerID="cri-o://d9a2ab24e902dcae258c6c3344acdb65b1fd26e52caba9b4a100dd420871cb26" gracePeriod=30 Nov 24 14:39:00 crc kubenswrapper[4822]: I1124 14:39:00.493075 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="sg-core" containerID="cri-o://86c4fcfee95aa28d4015bfda8accace4b928a4f937980afa151961aa0876fb91" gracePeriod=30 Nov 24 14:39:00 crc kubenswrapper[4822]: I1124 14:39:00.493131 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="ceilometer-notification-agent" containerID="cri-o://b6799e2d2ec8963b470f1896ff5c09f2582ee76336dc78bdb0be4273b1d43f6a" gracePeriod=30 Nov 24 14:39:00 crc kubenswrapper[4822]: I1124 14:39:00.523049 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 24 14:39:01 crc kubenswrapper[4822]: I1124 14:39:01.046710 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:39:01 crc kubenswrapper[4822]: E1124 14:39:01.047450 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:39:01 crc kubenswrapper[4822]: E1124 14:39:01.047736 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:39:01 crc kubenswrapper[4822]: E1124 14:39:01.047831 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:39:03.047815807 +0000 UTC m=+1180.164456284 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:39:01 crc kubenswrapper[4822]: I1124 14:39:01.174722 4822 generic.go:334] "Generic (PLEG): container finished" podID="be01d337-ad72-4656-82aa-33edea15526b" containerID="d9a2ab24e902dcae258c6c3344acdb65b1fd26e52caba9b4a100dd420871cb26" exitCode=0 Nov 24 14:39:01 crc kubenswrapper[4822]: I1124 14:39:01.174968 4822 generic.go:334] "Generic (PLEG): container finished" podID="be01d337-ad72-4656-82aa-33edea15526b" containerID="86c4fcfee95aa28d4015bfda8accace4b928a4f937980afa151961aa0876fb91" exitCode=2 Nov 24 14:39:01 crc kubenswrapper[4822]: I1124 14:39:01.175044 4822 generic.go:334] "Generic (PLEG): container finished" podID="be01d337-ad72-4656-82aa-33edea15526b" containerID="665a04c41c091fde310ae40cf0689f2e8e9ccabbc82fc659605155c48fda2444" exitCode=0 Nov 24 14:39:01 crc kubenswrapper[4822]: I1124 14:39:01.174789 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerDied","Data":"d9a2ab24e902dcae258c6c3344acdb65b1fd26e52caba9b4a100dd420871cb26"} Nov 24 14:39:01 crc kubenswrapper[4822]: I1124 14:39:01.175195 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerDied","Data":"86c4fcfee95aa28d4015bfda8accace4b928a4f937980afa151961aa0876fb91"} Nov 24 14:39:01 crc kubenswrapper[4822]: I1124 14:39:01.175682 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerDied","Data":"665a04c41c091fde310ae40cf0689f2e8e9ccabbc82fc659605155c48fda2444"} Nov 24 14:39:02 crc kubenswrapper[4822]: I1124 14:39:02.205510 4822 generic.go:334] "Generic (PLEG): container finished" podID="be01d337-ad72-4656-82aa-33edea15526b" containerID="b6799e2d2ec8963b470f1896ff5c09f2582ee76336dc78bdb0be4273b1d43f6a" exitCode=0 Nov 24 14:39:02 crc kubenswrapper[4822]: I1124 14:39:02.205602 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerDied","Data":"b6799e2d2ec8963b470f1896ff5c09f2582ee76336dc78bdb0be4273b1d43f6a"} Nov 24 14:39:03 crc kubenswrapper[4822]: I1124 14:39:03.091865 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:39:03 crc kubenswrapper[4822]: E1124 14:39:03.092151 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:39:03 crc kubenswrapper[4822]: E1124 14:39:03.092197 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:39:03 crc kubenswrapper[4822]: E1124 14:39:03.092319 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:39:07.092288497 +0000 UTC m=+1184.208929014 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.351191 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447076 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-run-httpd\") pod \"be01d337-ad72-4656-82aa-33edea15526b\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447154 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-sg-core-conf-yaml\") pod \"be01d337-ad72-4656-82aa-33edea15526b\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447178 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-config-data\") pod \"be01d337-ad72-4656-82aa-33edea15526b\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447241 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-log-httpd\") pod \"be01d337-ad72-4656-82aa-33edea15526b\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447259 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-scripts\") pod \"be01d337-ad72-4656-82aa-33edea15526b\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447305 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vrz9\" (UniqueName: \"kubernetes.io/projected/be01d337-ad72-4656-82aa-33edea15526b-kube-api-access-7vrz9\") pod \"be01d337-ad72-4656-82aa-33edea15526b\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447427 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-combined-ca-bundle\") pod \"be01d337-ad72-4656-82aa-33edea15526b\" (UID: \"be01d337-ad72-4656-82aa-33edea15526b\") " Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447619 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "be01d337-ad72-4656-82aa-33edea15526b" (UID: "be01d337-ad72-4656-82aa-33edea15526b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447940 4822 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.447951 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "be01d337-ad72-4656-82aa-33edea15526b" (UID: "be01d337-ad72-4656-82aa-33edea15526b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.452155 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-scripts" (OuterVolumeSpecName: "scripts") pod "be01d337-ad72-4656-82aa-33edea15526b" (UID: "be01d337-ad72-4656-82aa-33edea15526b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.452397 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be01d337-ad72-4656-82aa-33edea15526b-kube-api-access-7vrz9" (OuterVolumeSpecName: "kube-api-access-7vrz9") pod "be01d337-ad72-4656-82aa-33edea15526b" (UID: "be01d337-ad72-4656-82aa-33edea15526b"). InnerVolumeSpecName "kube-api-access-7vrz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.488242 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "be01d337-ad72-4656-82aa-33edea15526b" (UID: "be01d337-ad72-4656-82aa-33edea15526b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.525171 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be01d337-ad72-4656-82aa-33edea15526b" (UID: "be01d337-ad72-4656-82aa-33edea15526b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.550458 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.550504 4822 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.550519 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.550534 4822 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be01d337-ad72-4656-82aa-33edea15526b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.550561 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vrz9\" (UniqueName: \"kubernetes.io/projected/be01d337-ad72-4656-82aa-33edea15526b-kube-api-access-7vrz9\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.584815 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-config-data" (OuterVolumeSpecName: "config-data") pod "be01d337-ad72-4656-82aa-33edea15526b" (UID: "be01d337-ad72-4656-82aa-33edea15526b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4822]: I1124 14:39:05.655776 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be01d337-ad72-4656-82aa-33edea15526b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.003124 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.003595 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-log" containerID="cri-o://09bdf186b9cc111d712361299a44adf688cfdffe95daf750743a3bd73aa6ebc8" gracePeriod=30 Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.003727 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-httpd" containerID="cri-o://8fba0d1620c0c0b27290a0c8006fbcd806f65acec359a6e086370786aab000b1" gracePeriod=30 Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.171453 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.264962 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"15618ea1-3097-4bd0-aab1-8feedfc75839","Type":"ContainerStarted","Data":"dbcfa4951b86c74423f51f414762b7672381591c83e1c5584cdb5a603df71d5b"} Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.265781 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data\") pod \"a0aeb993-7d61-4924-bc31-86c00d8accb5\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.265824 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhq64\" (UniqueName: \"kubernetes.io/projected/a0aeb993-7d61-4924-bc31-86c00d8accb5-kube-api-access-hhq64\") pod \"a0aeb993-7d61-4924-bc31-86c00d8accb5\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.266059 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data-custom\") pod \"a0aeb993-7d61-4924-bc31-86c00d8accb5\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.266261 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-combined-ca-bundle\") pod \"a0aeb993-7d61-4924-bc31-86c00d8accb5\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.266324 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0aeb993-7d61-4924-bc31-86c00d8accb5-etc-machine-id\") pod \"a0aeb993-7d61-4924-bc31-86c00d8accb5\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.266475 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0aeb993-7d61-4924-bc31-86c00d8accb5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a0aeb993-7d61-4924-bc31-86c00d8accb5" (UID: "a0aeb993-7d61-4924-bc31-86c00d8accb5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.266867 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0aeb993-7d61-4924-bc31-86c00d8accb5-logs\") pod \"a0aeb993-7d61-4924-bc31-86c00d8accb5\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.266907 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-scripts\") pod \"a0aeb993-7d61-4924-bc31-86c00d8accb5\" (UID: \"a0aeb993-7d61-4924-bc31-86c00d8accb5\") " Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.267247 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0aeb993-7d61-4924-bc31-86c00d8accb5-logs" (OuterVolumeSpecName: "logs") pod "a0aeb993-7d61-4924-bc31-86c00d8accb5" (UID: "a0aeb993-7d61-4924-bc31-86c00d8accb5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.267652 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0aeb993-7d61-4924-bc31-86c00d8accb5-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.267673 4822 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a0aeb993-7d61-4924-bc31-86c00d8accb5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.273912 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-scripts" (OuterVolumeSpecName: "scripts") pod "a0aeb993-7d61-4924-bc31-86c00d8accb5" (UID: "a0aeb993-7d61-4924-bc31-86c00d8accb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.273949 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a0aeb993-7d61-4924-bc31-86c00d8accb5" (UID: "a0aeb993-7d61-4924-bc31-86c00d8accb5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.273967 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0aeb993-7d61-4924-bc31-86c00d8accb5-kube-api-access-hhq64" (OuterVolumeSpecName: "kube-api-access-hhq64") pod "a0aeb993-7d61-4924-bc31-86c00d8accb5" (UID: "a0aeb993-7d61-4924-bc31-86c00d8accb5"). InnerVolumeSpecName "kube-api-access-hhq64". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.303716 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be01d337-ad72-4656-82aa-33edea15526b","Type":"ContainerDied","Data":"44eebce83c63d97b358b67832009ae814922e304a696f7c05748f63057cae5b0"} Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.303788 4822 scope.go:117] "RemoveContainer" containerID="d9a2ab24e902dcae258c6c3344acdb65b1fd26e52caba9b4a100dd420871cb26" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.303955 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.326291 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0aeb993-7d61-4924-bc31-86c00d8accb5" (UID: "a0aeb993-7d61-4924-bc31-86c00d8accb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.331359 4822 generic.go:334] "Generic (PLEG): container finished" podID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerID="1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044" exitCode=137 Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.331543 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.331677 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a0aeb993-7d61-4924-bc31-86c00d8accb5","Type":"ContainerDied","Data":"1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044"} Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.331721 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a0aeb993-7d61-4924-bc31-86c00d8accb5","Type":"ContainerDied","Data":"efff0764f6c6d93fc8d1fa74d29a690714957ae90d21d29e67c4c068aa553df7"} Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.333965 4822 generic.go:334] "Generic (PLEG): container finished" podID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerID="09bdf186b9cc111d712361299a44adf688cfdffe95daf750743a3bd73aa6ebc8" exitCode=143 Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.334005 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ae3c17ab-662f-4507-a8c6-24a738cd3305","Type":"ContainerDied","Data":"09bdf186b9cc111d712361299a44adf688cfdffe95daf750743a3bd73aa6ebc8"} Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.370081 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.370111 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.370121 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhq64\" (UniqueName: \"kubernetes.io/projected/a0aeb993-7d61-4924-bc31-86c00d8accb5-kube-api-access-hhq64\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.370133 4822 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.371955 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data" (OuterVolumeSpecName: "config-data") pod "a0aeb993-7d61-4924-bc31-86c00d8accb5" (UID: "a0aeb993-7d61-4924-bc31-86c00d8accb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.386807 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.173753247 podStartE2EDuration="12.38679107s" podCreationTimestamp="2025-11-24 14:38:54 +0000 UTC" firstStartedPulling="2025-11-24 14:38:55.801958195 +0000 UTC m=+1172.918598672" lastFinishedPulling="2025-11-24 14:39:05.014996008 +0000 UTC m=+1182.131636495" observedRunningTime="2025-11-24 14:39:06.301288216 +0000 UTC m=+1183.417928693" watchObservedRunningTime="2025-11-24 14:39:06.38679107 +0000 UTC m=+1183.503431547" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.396095 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.405331 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.405470 4822 scope.go:117] "RemoveContainer" containerID="86c4fcfee95aa28d4015bfda8accace4b928a4f937980afa151961aa0876fb91" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.428960 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:06 crc kubenswrapper[4822]: E1124 14:39:06.429383 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerName="cinder-api" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429401 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerName="cinder-api" Nov 24 14:39:06 crc kubenswrapper[4822]: E1124 14:39:06.429422 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="ceilometer-notification-agent" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429428 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="ceilometer-notification-agent" Nov 24 14:39:06 crc kubenswrapper[4822]: E1124 14:39:06.429443 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="ceilometer-central-agent" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429449 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="ceilometer-central-agent" Nov 24 14:39:06 crc kubenswrapper[4822]: E1124 14:39:06.429458 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="sg-core" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429464 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="sg-core" Nov 24 14:39:06 crc kubenswrapper[4822]: E1124 14:39:06.429487 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="proxy-httpd" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429492 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="proxy-httpd" Nov 24 14:39:06 crc kubenswrapper[4822]: E1124 14:39:06.429506 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerName="cinder-api-log" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429512 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerName="cinder-api-log" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429690 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="proxy-httpd" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429706 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerName="cinder-api" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429717 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="sg-core" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429726 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="ceilometer-notification-agent" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429740 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="be01d337-ad72-4656-82aa-33edea15526b" containerName="ceilometer-central-agent" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.429753 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0aeb993-7d61-4924-bc31-86c00d8accb5" containerName="cinder-api-log" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.431537 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.433867 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.434219 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.444663 4822 scope.go:117] "RemoveContainer" containerID="b6799e2d2ec8963b470f1896ff5c09f2582ee76336dc78bdb0be4273b1d43f6a" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.455925 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.472531 4822 scope.go:117] "RemoveContainer" containerID="665a04c41c091fde310ae40cf0689f2e8e9ccabbc82fc659605155c48fda2444" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.472719 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0aeb993-7d61-4924-bc31-86c00d8accb5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.498781 4822 scope.go:117] "RemoveContainer" containerID="1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.527918 4822 scope.go:117] "RemoveContainer" containerID="ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.546537 4822 scope.go:117] "RemoveContainer" containerID="1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044" Nov 24 14:39:06 crc kubenswrapper[4822]: E1124 14:39:06.546981 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044\": container with ID starting with 1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044 not found: ID does not exist" containerID="1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.547043 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044"} err="failed to get container status \"1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044\": rpc error: code = NotFound desc = could not find container \"1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044\": container with ID starting with 1ca83a4251cb3c64ac02c6009a16531dc0fae651c313e1178ace619e0e5db044 not found: ID does not exist" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.547075 4822 scope.go:117] "RemoveContainer" containerID="ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c" Nov 24 14:39:06 crc kubenswrapper[4822]: E1124 14:39:06.547372 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c\": container with ID starting with ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c not found: ID does not exist" containerID="ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.547405 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c"} err="failed to get container status \"ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c\": rpc error: code = NotFound desc = could not find container \"ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c\": container with ID starting with ee571c172f2091d6cc801ed180e5d8e22cd92303a969361c1e1634bbc2ff9e9c not found: ID does not exist" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.575012 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.575099 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-scripts\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.575161 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-run-httpd\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.575182 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.575311 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-config-data\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.575341 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48s5z\" (UniqueName: \"kubernetes.io/projected/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-kube-api-access-48s5z\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.575406 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-log-httpd\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.663713 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.671306 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.677140 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-run-httpd\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.677197 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.677264 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-config-data\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.677293 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48s5z\" (UniqueName: \"kubernetes.io/projected/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-kube-api-access-48s5z\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.677325 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-log-httpd\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.677437 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.677485 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-scripts\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.677667 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-run-httpd\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.677815 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-log-httpd\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.680957 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.691668 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.692355 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-scripts\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.695121 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.703714 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.707230 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.707668 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.707918 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.709784 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-config-data\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.711671 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48s5z\" (UniqueName: \"kubernetes.io/projected/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-kube-api-access-48s5z\") pod \"ceilometer-0\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.734600 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.753033 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.778951 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmwmh\" (UniqueName: \"kubernetes.io/projected/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-kube-api-access-tmwmh\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.778994 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.779082 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-config-data\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.779109 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-config-data-custom\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.779127 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-scripts\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.779163 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-logs\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.779195 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.779243 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.779272 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.881287 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-config-data-custom\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.881549 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-scripts\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.881582 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-logs\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.881621 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.881654 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.881684 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.881741 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmwmh\" (UniqueName: \"kubernetes.io/projected/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-kube-api-access-tmwmh\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.881758 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.881817 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-config-data\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.888273 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-config-data\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.902478 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-config-data-custom\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.902561 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.906717 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.907506 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.907687 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-logs\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.908777 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.910595 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-scripts\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:06 crc kubenswrapper[4822]: I1124 14:39:06.938228 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmwmh\" (UniqueName: \"kubernetes.io/projected/5c756c92-6b13-4f1d-b3c0-b217c926a2c8-kube-api-access-tmwmh\") pod \"cinder-api-0\" (UID: \"5c756c92-6b13-4f1d-b3c0-b217c926a2c8\") " pod="openstack/cinder-api-0" Nov 24 14:39:07 crc kubenswrapper[4822]: I1124 14:39:07.036809 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:39:07 crc kubenswrapper[4822]: I1124 14:39:07.188471 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:39:07 crc kubenswrapper[4822]: E1124 14:39:07.189353 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:39:07 crc kubenswrapper[4822]: E1124 14:39:07.189372 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:39:07 crc kubenswrapper[4822]: E1124 14:39:07.189419 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:39:15.189403077 +0000 UTC m=+1192.306043554 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:39:07 crc kubenswrapper[4822]: I1124 14:39:07.331372 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:07 crc kubenswrapper[4822]: I1124 14:39:07.346510 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:39:07 crc kubenswrapper[4822]: I1124 14:39:07.539145 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:39:07 crc kubenswrapper[4822]: W1124 14:39:07.544440 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c756c92_6b13_4f1d_b3c0_b217c926a2c8.slice/crio-8575b82c94c0c862b30e70aaf7e0f5bddd01283d2c76890d2f685058e87db142 WatchSource:0}: Error finding container 8575b82c94c0c862b30e70aaf7e0f5bddd01283d2c76890d2f685058e87db142: Status 404 returned error can't find the container with id 8575b82c94c0c862b30e70aaf7e0f5bddd01283d2c76890d2f685058e87db142 Nov 24 14:39:07 crc kubenswrapper[4822]: I1124 14:39:07.730743 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0aeb993-7d61-4924-bc31-86c00d8accb5" path="/var/lib/kubelet/pods/a0aeb993-7d61-4924-bc31-86c00d8accb5/volumes" Nov 24 14:39:07 crc kubenswrapper[4822]: I1124 14:39:07.732363 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be01d337-ad72-4656-82aa-33edea15526b" path="/var/lib/kubelet/pods/be01d337-ad72-4656-82aa-33edea15526b/volumes" Nov 24 14:39:08 crc kubenswrapper[4822]: I1124 14:39:08.369905 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerStarted","Data":"4c5edc72161a570b1de033ee7b676a4e7e5aaaaea501331e1c8a81887e6f64ca"} Nov 24 14:39:08 crc kubenswrapper[4822]: I1124 14:39:08.370571 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerStarted","Data":"0a0c8da07d8e101b489d2c53645d3abafe8bddc6a32ae7b1442f48aec01c29c7"} Nov 24 14:39:08 crc kubenswrapper[4822]: I1124 14:39:08.371868 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5c756c92-6b13-4f1d-b3c0-b217c926a2c8","Type":"ContainerStarted","Data":"8a1e7a7267c7eb7a9d10124c52a1e939daf6fec70c733154da4b0556876e1fac"} Nov 24 14:39:08 crc kubenswrapper[4822]: I1124 14:39:08.371931 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5c756c92-6b13-4f1d-b3c0-b217c926a2c8","Type":"ContainerStarted","Data":"8575b82c94c0c862b30e70aaf7e0f5bddd01283d2c76890d2f685058e87db142"} Nov 24 14:39:08 crc kubenswrapper[4822]: I1124 14:39:08.908525 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:09 crc kubenswrapper[4822]: I1124 14:39:09.401271 4822 generic.go:334] "Generic (PLEG): container finished" podID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerID="8fba0d1620c0c0b27290a0c8006fbcd806f65acec359a6e086370786aab000b1" exitCode=0 Nov 24 14:39:09 crc kubenswrapper[4822]: I1124 14:39:09.401306 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ae3c17ab-662f-4507-a8c6-24a738cd3305","Type":"ContainerDied","Data":"8fba0d1620c0c0b27290a0c8006fbcd806f65acec359a6e086370786aab000b1"} Nov 24 14:39:09 crc kubenswrapper[4822]: I1124 14:39:09.403145 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerStarted","Data":"91dc843e037f676c81fcb7a6d1b291edf12bf9e5f902db6f442d93636a5ff089"} Nov 24 14:39:09 crc kubenswrapper[4822]: I1124 14:39:09.404731 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5c756c92-6b13-4f1d-b3c0-b217c926a2c8","Type":"ContainerStarted","Data":"b145c8c924e5441a85553bb6d1444b660178a4141c1ad4baaede1ae68494145c"} Nov 24 14:39:09 crc kubenswrapper[4822]: I1124 14:39:09.404912 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 14:39:09 crc kubenswrapper[4822]: I1124 14:39:09.423522 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.423505479 podStartE2EDuration="3.423505479s" podCreationTimestamp="2025-11-24 14:39:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:09.419550932 +0000 UTC m=+1186.536191409" watchObservedRunningTime="2025-11-24 14:39:09.423505479 +0000 UTC m=+1186.540145956" Nov 24 14:39:10 crc kubenswrapper[4822]: I1124 14:39:10.671325 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.167:9292/healthcheck\": dial tcp 10.217.0.167:9292: connect: connection refused" Nov 24 14:39:10 crc kubenswrapper[4822]: I1124 14:39:10.671396 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.167:9292/healthcheck\": dial tcp 10.217.0.167:9292: connect: connection refused" Nov 24 14:39:10 crc kubenswrapper[4822]: I1124 14:39:10.815755 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:39:10 crc kubenswrapper[4822]: I1124 14:39:10.816001 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerName="glance-log" containerID="cri-o://cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e" gracePeriod=30 Nov 24 14:39:10 crc kubenswrapper[4822]: I1124 14:39:10.816362 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerName="glance-httpd" containerID="cri-o://6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06" gracePeriod=30 Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.325315 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.382119 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25q7b\" (UniqueName: \"kubernetes.io/projected/ae3c17ab-662f-4507-a8c6-24a738cd3305-kube-api-access-25q7b\") pod \"ae3c17ab-662f-4507-a8c6-24a738cd3305\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.382389 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-httpd-run\") pod \"ae3c17ab-662f-4507-a8c6-24a738cd3305\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.382466 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-public-tls-certs\") pod \"ae3c17ab-662f-4507-a8c6-24a738cd3305\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.382847 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ae3c17ab-662f-4507-a8c6-24a738cd3305" (UID: "ae3c17ab-662f-4507-a8c6-24a738cd3305"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.383094 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-config-data\") pod \"ae3c17ab-662f-4507-a8c6-24a738cd3305\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.383192 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"ae3c17ab-662f-4507-a8c6-24a738cd3305\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.383234 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-logs\") pod \"ae3c17ab-662f-4507-a8c6-24a738cd3305\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.383293 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-scripts\") pod \"ae3c17ab-662f-4507-a8c6-24a738cd3305\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.383330 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-combined-ca-bundle\") pod \"ae3c17ab-662f-4507-a8c6-24a738cd3305\" (UID: \"ae3c17ab-662f-4507-a8c6-24a738cd3305\") " Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.383814 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-logs" (OuterVolumeSpecName: "logs") pod "ae3c17ab-662f-4507-a8c6-24a738cd3305" (UID: "ae3c17ab-662f-4507-a8c6-24a738cd3305"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.383873 4822 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.394293 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-scripts" (OuterVolumeSpecName: "scripts") pod "ae3c17ab-662f-4507-a8c6-24a738cd3305" (UID: "ae3c17ab-662f-4507-a8c6-24a738cd3305"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.403088 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae3c17ab-662f-4507-a8c6-24a738cd3305-kube-api-access-25q7b" (OuterVolumeSpecName: "kube-api-access-25q7b") pod "ae3c17ab-662f-4507-a8c6-24a738cd3305" (UID: "ae3c17ab-662f-4507-a8c6-24a738cd3305"). InnerVolumeSpecName "kube-api-access-25q7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.431450 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1" (OuterVolumeSpecName: "glance") pod "ae3c17ab-662f-4507-a8c6-24a738cd3305" (UID: "ae3c17ab-662f-4507-a8c6-24a738cd3305"). InnerVolumeSpecName "pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.450912 4822 generic.go:334] "Generic (PLEG): container finished" podID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerID="cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e" exitCode=143 Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.450981 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb16d0ea-2a9a-4b31-897c-eb6938531d2e","Type":"ContainerDied","Data":"cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e"} Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.487479 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.487521 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25q7b\" (UniqueName: \"kubernetes.io/projected/ae3c17ab-662f-4507-a8c6-24a738cd3305-kube-api-access-25q7b\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.487546 4822 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") on node \"crc\" " Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.487557 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae3c17ab-662f-4507-a8c6-24a738cd3305-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.496568 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.497424 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ae3c17ab-662f-4507-a8c6-24a738cd3305","Type":"ContainerDied","Data":"02af026aa88c2847acb335ad83d45fddebf5b1429c38def0ca42fc61eae931e7"} Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.497469 4822 scope.go:117] "RemoveContainer" containerID="8fba0d1620c0c0b27290a0c8006fbcd806f65acec359a6e086370786aab000b1" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.511562 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerStarted","Data":"efe18120e9205b524c22f4cb8078b8a34ff3231d0f7b13943c65d11c0fdf1695"} Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.512431 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae3c17ab-662f-4507-a8c6-24a738cd3305" (UID: "ae3c17ab-662f-4507-a8c6-24a738cd3305"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.537458 4822 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.537618 4822 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1") on node "crc" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.539331 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-config-data" (OuterVolumeSpecName: "config-data") pod "ae3c17ab-662f-4507-a8c6-24a738cd3305" (UID: "ae3c17ab-662f-4507-a8c6-24a738cd3305"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.550515 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ae3c17ab-662f-4507-a8c6-24a738cd3305" (UID: "ae3c17ab-662f-4507-a8c6-24a738cd3305"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.565771 4822 scope.go:117] "RemoveContainer" containerID="09bdf186b9cc111d712361299a44adf688cfdffe95daf750743a3bd73aa6ebc8" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.590812 4822 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.590846 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.590857 4822 reconciler_common.go:293] "Volume detached for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.590869 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3c17ab-662f-4507-a8c6-24a738cd3305-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.825597 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.835217 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.852885 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:39:11 crc kubenswrapper[4822]: E1124 14:39:11.853355 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-log" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.853375 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-log" Nov 24 14:39:11 crc kubenswrapper[4822]: E1124 14:39:11.853402 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-httpd" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.853408 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-httpd" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.853584 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-httpd" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.853611 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" containerName="glance-log" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.854702 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.858339 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.858526 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.863526 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.997367 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.997424 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-config-data\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.997449 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.997489 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.997511 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcls2\" (UniqueName: \"kubernetes.io/projected/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-kube-api-access-rcls2\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.997545 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-logs\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.997599 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-scripts\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:11 crc kubenswrapper[4822]: I1124 14:39:11.997625 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.099985 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.100046 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-config-data\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.100072 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.100109 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.100130 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcls2\" (UniqueName: \"kubernetes.io/projected/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-kube-api-access-rcls2\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.100167 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-logs\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.100213 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-scripts\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.100241 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.101790 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-logs\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.101794 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.106866 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-scripts\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.107615 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.107876 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-config-data\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.108437 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.116797 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.116838 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2ffde4dda1be13a8596eb6948dbe2a502fe570389c8cc35530979ae06145ae21/globalmount\"" pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.127250 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcls2\" (UniqueName: \"kubernetes.io/projected/e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9-kube-api-access-rcls2\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.167474 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f23b7bab-becc-4f10-8bca-1123d9ac62d1\") pod \"glance-default-external-api-0\" (UID: \"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9\") " pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.192835 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:39:12 crc kubenswrapper[4822]: W1124 14:39:12.737302 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0e7dc25_3ad1_4deb_88a0_5522ae9e6da9.slice/crio-8d953c6ccbbb2973f5d41e82dfb3f454c0b684967a425b72252d429927d41fed WatchSource:0}: Error finding container 8d953c6ccbbb2973f5d41e82dfb3f454c0b684967a425b72252d429927d41fed: Status 404 returned error can't find the container with id 8d953c6ccbbb2973f5d41e82dfb3f454c0b684967a425b72252d429927d41fed Nov 24 14:39:12 crc kubenswrapper[4822]: I1124 14:39:12.744051 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.555979 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9","Type":"ContainerStarted","Data":"90bb628f45a3d845ecca4f79f7c3b01977b83f4a7141914c2969521e39710a73"} Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.556466 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9","Type":"ContainerStarted","Data":"8d953c6ccbbb2973f5d41e82dfb3f454c0b684967a425b72252d429927d41fed"} Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.566447 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerStarted","Data":"88091be5bc420884d346f15745261462973450a471f0bc4cf4066bf19cb2533e"} Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.566629 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="ceilometer-central-agent" containerID="cri-o://4c5edc72161a570b1de033ee7b676a4e7e5aaaaea501331e1c8a81887e6f64ca" gracePeriod=30 Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.566886 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.567171 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="proxy-httpd" containerID="cri-o://88091be5bc420884d346f15745261462973450a471f0bc4cf4066bf19cb2533e" gracePeriod=30 Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.567234 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="sg-core" containerID="cri-o://efe18120e9205b524c22f4cb8078b8a34ff3231d0f7b13943c65d11c0fdf1695" gracePeriod=30 Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.567273 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="ceilometer-notification-agent" containerID="cri-o://91dc843e037f676c81fcb7a6d1b291edf12bf9e5f902db6f442d93636a5ff089" gracePeriod=30 Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.605319 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.014546007 podStartE2EDuration="7.60528786s" podCreationTimestamp="2025-11-24 14:39:06 +0000 UTC" firstStartedPulling="2025-11-24 14:39:07.346315201 +0000 UTC m=+1184.462955678" lastFinishedPulling="2025-11-24 14:39:12.937057054 +0000 UTC m=+1190.053697531" observedRunningTime="2025-11-24 14:39:13.596726788 +0000 UTC m=+1190.713367265" watchObservedRunningTime="2025-11-24 14:39:13.60528786 +0000 UTC m=+1190.721928337" Nov 24 14:39:13 crc kubenswrapper[4822]: I1124 14:39:13.743941 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae3c17ab-662f-4507-a8c6-24a738cd3305" path="/var/lib/kubelet/pods/ae3c17ab-662f-4507-a8c6-24a738cd3305/volumes" Nov 24 14:39:13 crc kubenswrapper[4822]: E1124 14:39:13.832435 4822 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod360c143b_2fba_4599_b6ff_5ee4a4cb32f6.slice/crio-efe18120e9205b524c22f4cb8078b8a34ff3231d0f7b13943c65d11c0fdf1695.scope\": RecentStats: unable to find data in memory cache]" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.542751 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.593766 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9","Type":"ContainerStarted","Data":"53500237e0ed29b5280d45e09e300ec8eafb80a173148545833a93777297c8db"} Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.601977 4822 generic.go:334] "Generic (PLEG): container finished" podID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerID="6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06" exitCode=0 Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.602159 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb16d0ea-2a9a-4b31-897c-eb6938531d2e","Type":"ContainerDied","Data":"6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06"} Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.602256 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb16d0ea-2a9a-4b31-897c-eb6938531d2e","Type":"ContainerDied","Data":"5c501dccd586a4ed774543deb55537bc98e6b79be183d21e4a32fb96690a62d0"} Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.602402 4822 scope.go:117] "RemoveContainer" containerID="6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.602570 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.625873 4822 generic.go:334] "Generic (PLEG): container finished" podID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerID="88091be5bc420884d346f15745261462973450a471f0bc4cf4066bf19cb2533e" exitCode=0 Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.626068 4822 generic.go:334] "Generic (PLEG): container finished" podID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerID="efe18120e9205b524c22f4cb8078b8a34ff3231d0f7b13943c65d11c0fdf1695" exitCode=2 Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.626123 4822 generic.go:334] "Generic (PLEG): container finished" podID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerID="91dc843e037f676c81fcb7a6d1b291edf12bf9e5f902db6f442d93636a5ff089" exitCode=0 Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.626228 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerDied","Data":"88091be5bc420884d346f15745261462973450a471f0bc4cf4066bf19cb2533e"} Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.626310 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerDied","Data":"efe18120e9205b524c22f4cb8078b8a34ff3231d0f7b13943c65d11c0fdf1695"} Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.626375 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerDied","Data":"91dc843e037f676c81fcb7a6d1b291edf12bf9e5f902db6f442d93636a5ff089"} Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.637335 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.637305671 podStartE2EDuration="3.637305671s" podCreationTimestamp="2025-11-24 14:39:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:14.620178626 +0000 UTC m=+1191.736819133" watchObservedRunningTime="2025-11-24 14:39:14.637305671 +0000 UTC m=+1191.753946148" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.652167 4822 scope.go:117] "RemoveContainer" containerID="cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.657960 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-httpd-run\") pod \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.658016 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-scripts\") pod \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.658096 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-combined-ca-bundle\") pod \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.658124 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-config-data\") pod \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.658349 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.658400 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw56s\" (UniqueName: \"kubernetes.io/projected/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-kube-api-access-fw56s\") pod \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.658436 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-internal-tls-certs\") pod \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.658478 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-logs\") pod \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\" (UID: \"fb16d0ea-2a9a-4b31-897c-eb6938531d2e\") " Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.658740 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fb16d0ea-2a9a-4b31-897c-eb6938531d2e" (UID: "fb16d0ea-2a9a-4b31-897c-eb6938531d2e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.659021 4822 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.661373 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-logs" (OuterVolumeSpecName: "logs") pod "fb16d0ea-2a9a-4b31-897c-eb6938531d2e" (UID: "fb16d0ea-2a9a-4b31-897c-eb6938531d2e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.664425 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-scripts" (OuterVolumeSpecName: "scripts") pod "fb16d0ea-2a9a-4b31-897c-eb6938531d2e" (UID: "fb16d0ea-2a9a-4b31-897c-eb6938531d2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.668420 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-kube-api-access-fw56s" (OuterVolumeSpecName: "kube-api-access-fw56s") pod "fb16d0ea-2a9a-4b31-897c-eb6938531d2e" (UID: "fb16d0ea-2a9a-4b31-897c-eb6938531d2e"). InnerVolumeSpecName "kube-api-access-fw56s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.677157 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0" (OuterVolumeSpecName: "glance") pod "fb16d0ea-2a9a-4b31-897c-eb6938531d2e" (UID: "fb16d0ea-2a9a-4b31-897c-eb6938531d2e"). InnerVolumeSpecName "pvc-49360db5-9984-4b54-aac7-d578276794c0". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.689023 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb16d0ea-2a9a-4b31-897c-eb6938531d2e" (UID: "fb16d0ea-2a9a-4b31-897c-eb6938531d2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.701860 4822 scope.go:117] "RemoveContainer" containerID="6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06" Nov 24 14:39:14 crc kubenswrapper[4822]: E1124 14:39:14.703433 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06\": container with ID starting with 6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06 not found: ID does not exist" containerID="6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.703487 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06"} err="failed to get container status \"6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06\": rpc error: code = NotFound desc = could not find container \"6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06\": container with ID starting with 6113c57604e91bc98cc8d786016e7b1f78e7412369e837f053efbf37a70c1e06 not found: ID does not exist" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.703516 4822 scope.go:117] "RemoveContainer" containerID="cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e" Nov 24 14:39:14 crc kubenswrapper[4822]: E1124 14:39:14.706369 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e\": container with ID starting with cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e not found: ID does not exist" containerID="cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.706418 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e"} err="failed to get container status \"cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e\": rpc error: code = NotFound desc = could not find container \"cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e\": container with ID starting with cdf64dc532cd95b3e3f2f275e50f9ad92cb17f3b81f51d1b7c37dd92c850f97e not found: ID does not exist" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.720434 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-config-data" (OuterVolumeSpecName: "config-data") pod "fb16d0ea-2a9a-4b31-897c-eb6938531d2e" (UID: "fb16d0ea-2a9a-4b31-897c-eb6938531d2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.738305 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fb16d0ea-2a9a-4b31-897c-eb6938531d2e" (UID: "fb16d0ea-2a9a-4b31-897c-eb6938531d2e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.762360 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.763060 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.763096 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.763129 4822 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") on node \"crc\" " Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.763144 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw56s\" (UniqueName: \"kubernetes.io/projected/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-kube-api-access-fw56s\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.763160 4822 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.763172 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb16d0ea-2a9a-4b31-897c-eb6938531d2e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.796378 4822 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.796537 4822 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-49360db5-9984-4b54-aac7-d578276794c0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0") on node "crc" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.865156 4822 reconciler_common.go:293] "Volume detached for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.977013 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:39:14 crc kubenswrapper[4822]: I1124 14:39:14.995301 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.026696 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:39:15 crc kubenswrapper[4822]: E1124 14:39:15.027173 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerName="glance-log" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.027216 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerName="glance-log" Nov 24 14:39:15 crc kubenswrapper[4822]: E1124 14:39:15.027249 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerName="glance-httpd" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.027255 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerName="glance-httpd" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.027436 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerName="glance-httpd" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.027455 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" containerName="glance-log" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.028630 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.031592 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.031816 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.043446 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.070627 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.070702 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.070779 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.070805 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.070822 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33a75f4c-1afc-43a0-81cb-2662d187cde4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.070841 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a75f4c-1afc-43a0-81cb-2662d187cde4-logs\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.070906 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.070928 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd89v\" (UniqueName: \"kubernetes.io/projected/33a75f4c-1afc-43a0-81cb-2662d187cde4-kube-api-access-cd89v\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.172411 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.172467 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33a75f4c-1afc-43a0-81cb-2662d187cde4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.172490 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a75f4c-1afc-43a0-81cb-2662d187cde4-logs\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.172549 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.172570 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd89v\" (UniqueName: \"kubernetes.io/projected/33a75f4c-1afc-43a0-81cb-2662d187cde4-kube-api-access-cd89v\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.172634 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.172663 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.172722 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.174516 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33a75f4c-1afc-43a0-81cb-2662d187cde4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.174707 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a75f4c-1afc-43a0-81cb-2662d187cde4-logs\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.175571 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.175600 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4ad6c7c6bc95d7d12c286d711c7d508fc5826003cdc7d089436fe350e33badd7/globalmount\"" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.176631 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.176879 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.177858 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.184765 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33a75f4c-1afc-43a0-81cb-2662d187cde4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.200425 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd89v\" (UniqueName: \"kubernetes.io/projected/33a75f4c-1afc-43a0-81cb-2662d187cde4-kube-api-access-cd89v\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.248971 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-49360db5-9984-4b54-aac7-d578276794c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49360db5-9984-4b54-aac7-d578276794c0\") pod \"glance-default-internal-api-0\" (UID: \"33a75f4c-1afc-43a0-81cb-2662d187cde4\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.274659 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:39:15 crc kubenswrapper[4822]: E1124 14:39:15.274859 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:39:15 crc kubenswrapper[4822]: E1124 14:39:15.275176 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:39:15 crc kubenswrapper[4822]: E1124 14:39:15.275359 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:39:31.275333167 +0000 UTC m=+1208.391973644 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.352858 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.637348 4822 generic.go:334] "Generic (PLEG): container finished" podID="cfb584d1-8a63-4397-b137-65ac59e8808e" containerID="98b3025eb88fe0e6ca8ce2e7f0b15fc10da312d8940e44f3b4ba5e3bbaf522ef" exitCode=137 Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.637642 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"cfb584d1-8a63-4397-b137-65ac59e8808e","Type":"ContainerDied","Data":"98b3025eb88fe0e6ca8ce2e7f0b15fc10da312d8940e44f3b4ba5e3bbaf522ef"} Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.723018 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb16d0ea-2a9a-4b31-897c-eb6938531d2e" path="/var/lib/kubelet/pods/fb16d0ea-2a9a-4b31-897c-eb6938531d2e/volumes" Nov 24 14:39:15 crc kubenswrapper[4822]: I1124 14:39:15.948889 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:39:15 crc kubenswrapper[4822]: W1124 14:39:15.951393 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33a75f4c_1afc_43a0_81cb_2662d187cde4.slice/crio-29b9baef109340a5d35c0f9ca44ecd13eecfbf1fe59fb65d6060beb6f277b862 WatchSource:0}: Error finding container 29b9baef109340a5d35c0f9ca44ecd13eecfbf1fe59fb65d6060beb6f277b862: Status 404 returned error can't find the container with id 29b9baef109340a5d35c0f9ca44ecd13eecfbf1fe59fb65d6060beb6f277b862 Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.066005 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.203835 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-certs\") pod \"cfb584d1-8a63-4397-b137-65ac59e8808e\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.203909 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-scripts\") pod \"cfb584d1-8a63-4397-b137-65ac59e8808e\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.203951 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zrqm\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-kube-api-access-7zrqm\") pod \"cfb584d1-8a63-4397-b137-65ac59e8808e\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.204079 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-combined-ca-bundle\") pod \"cfb584d1-8a63-4397-b137-65ac59e8808e\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.204171 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data-custom\") pod \"cfb584d1-8a63-4397-b137-65ac59e8808e\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.204244 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data\") pod \"cfb584d1-8a63-4397-b137-65ac59e8808e\" (UID: \"cfb584d1-8a63-4397-b137-65ac59e8808e\") " Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.211197 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-certs" (OuterVolumeSpecName: "certs") pod "cfb584d1-8a63-4397-b137-65ac59e8808e" (UID: "cfb584d1-8a63-4397-b137-65ac59e8808e"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.211693 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-kube-api-access-7zrqm" (OuterVolumeSpecName: "kube-api-access-7zrqm") pod "cfb584d1-8a63-4397-b137-65ac59e8808e" (UID: "cfb584d1-8a63-4397-b137-65ac59e8808e"). InnerVolumeSpecName "kube-api-access-7zrqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.211763 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-scripts" (OuterVolumeSpecName: "scripts") pod "cfb584d1-8a63-4397-b137-65ac59e8808e" (UID: "cfb584d1-8a63-4397-b137-65ac59e8808e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.217285 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cfb584d1-8a63-4397-b137-65ac59e8808e" (UID: "cfb584d1-8a63-4397-b137-65ac59e8808e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.238598 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data" (OuterVolumeSpecName: "config-data") pod "cfb584d1-8a63-4397-b137-65ac59e8808e" (UID: "cfb584d1-8a63-4397-b137-65ac59e8808e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.248543 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfb584d1-8a63-4397-b137-65ac59e8808e" (UID: "cfb584d1-8a63-4397-b137-65ac59e8808e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.306887 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.306918 4822 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.306928 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.306936 4822 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.306944 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfb584d1-8a63-4397-b137-65ac59e8808e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.306951 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zrqm\" (UniqueName: \"kubernetes.io/projected/cfb584d1-8a63-4397-b137-65ac59e8808e-kube-api-access-7zrqm\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.648486 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"cfb584d1-8a63-4397-b137-65ac59e8808e","Type":"ContainerDied","Data":"519e0c7ab52a9e6b186e34f4edd7cd2c1dd528aaf29c4367a61e62ebda81862b"} Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.648737 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.648778 4822 scope.go:117] "RemoveContainer" containerID="98b3025eb88fe0e6ca8ce2e7f0b15fc10da312d8940e44f3b4ba5e3bbaf522ef" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.651939 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33a75f4c-1afc-43a0-81cb-2662d187cde4","Type":"ContainerStarted","Data":"662d02a71c14a0db54263b3e0d214efbfbf9e0f1b0627625d815c7e566519497"} Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.652006 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33a75f4c-1afc-43a0-81cb-2662d187cde4","Type":"ContainerStarted","Data":"29b9baef109340a5d35c0f9ca44ecd13eecfbf1fe59fb65d6060beb6f277b862"} Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.680632 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.701747 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.721581 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:39:16 crc kubenswrapper[4822]: E1124 14:39:16.722004 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb584d1-8a63-4397-b137-65ac59e8808e" containerName="cloudkitty-proc" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.722020 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb584d1-8a63-4397-b137-65ac59e8808e" containerName="cloudkitty-proc" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.722246 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb584d1-8a63-4397-b137-65ac59e8808e" containerName="cloudkitty-proc" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.722980 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.739081 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.752309 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.817693 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.817739 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.817831 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-certs\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.817875 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkgw7\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-kube-api-access-tkgw7\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.817899 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.818070 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-scripts\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.920094 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-scripts\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.920226 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.920257 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.920330 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-certs\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.920374 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkgw7\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-kube-api-access-tkgw7\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.920404 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.924846 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-certs\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.924840 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-scripts\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.925881 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.926468 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.930725 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:16 crc kubenswrapper[4822]: I1124 14:39:16.944921 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkgw7\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-kube-api-access-tkgw7\") pod \"cloudkitty-proc-0\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:39:17 crc kubenswrapper[4822]: I1124 14:39:17.106730 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:39:17 crc kubenswrapper[4822]: I1124 14:39:17.663630 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33a75f4c-1afc-43a0-81cb-2662d187cde4","Type":"ContainerStarted","Data":"36fd1d2a9d78a0d600ee37b73ffc2e9e5dbaddc09277c288d7eb1be7dd34ca7b"} Nov 24 14:39:17 crc kubenswrapper[4822]: I1124 14:39:17.691342 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.69132312 podStartE2EDuration="3.69132312s" podCreationTimestamp="2025-11-24 14:39:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:17.684759572 +0000 UTC m=+1194.801400049" watchObservedRunningTime="2025-11-24 14:39:17.69132312 +0000 UTC m=+1194.807963597" Nov 24 14:39:17 crc kubenswrapper[4822]: W1124 14:39:17.711956 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74794267_7a2e_4dc4_bb6b_35b99814e3ee.slice/crio-ac2d57e7f707b8d3babbd443fd878be465036b914de6d34d52ff777ad9e9d536 WatchSource:0}: Error finding container ac2d57e7f707b8d3babbd443fd878be465036b914de6d34d52ff777ad9e9d536: Status 404 returned error can't find the container with id ac2d57e7f707b8d3babbd443fd878be465036b914de6d34d52ff777ad9e9d536 Nov 24 14:39:17 crc kubenswrapper[4822]: I1124 14:39:17.716047 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfb584d1-8a63-4397-b137-65ac59e8808e" path="/var/lib/kubelet/pods/cfb584d1-8a63-4397-b137-65ac59e8808e/volumes" Nov 24 14:39:17 crc kubenswrapper[4822]: I1124 14:39:17.716896 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:39:18 crc kubenswrapper[4822]: I1124 14:39:18.675808 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"74794267-7a2e-4dc4-bb6b-35b99814e3ee","Type":"ContainerStarted","Data":"48c41eaf640975344f69e7f1b14f82dce5425912c516ece0728ca67e1715543d"} Nov 24 14:39:18 crc kubenswrapper[4822]: I1124 14:39:18.676133 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"74794267-7a2e-4dc4-bb6b-35b99814e3ee","Type":"ContainerStarted","Data":"ac2d57e7f707b8d3babbd443fd878be465036b914de6d34d52ff777ad9e9d536"} Nov 24 14:39:18 crc kubenswrapper[4822]: I1124 14:39:18.695353 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=2.6953303589999997 podStartE2EDuration="2.695330359s" podCreationTimestamp="2025-11-24 14:39:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:18.689039429 +0000 UTC m=+1195.805679916" watchObservedRunningTime="2025-11-24 14:39:18.695330359 +0000 UTC m=+1195.811970836" Nov 24 14:39:19 crc kubenswrapper[4822]: I1124 14:39:19.183361 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 14:39:21 crc kubenswrapper[4822]: I1124 14:39:21.705913 4822 generic.go:334] "Generic (PLEG): container finished" podID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerID="4c5edc72161a570b1de033ee7b676a4e7e5aaaaea501331e1c8a81887e6f64ca" exitCode=0 Nov 24 14:39:21 crc kubenswrapper[4822]: I1124 14:39:21.730990 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerDied","Data":"4c5edc72161a570b1de033ee7b676a4e7e5aaaaea501331e1c8a81887e6f64ca"} Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.166034 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.193776 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.193997 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.228987 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-combined-ca-bundle\") pod \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.229035 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-run-httpd\") pod \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.229066 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-config-data\") pod \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.229068 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.229157 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-log-httpd\") pod \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.229181 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-scripts\") pod \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.229305 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-sg-core-conf-yaml\") pod \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.229334 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48s5z\" (UniqueName: \"kubernetes.io/projected/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-kube-api-access-48s5z\") pod \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\" (UID: \"360c143b-2fba-4599-b6ff-5ee4a4cb32f6\") " Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.231662 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "360c143b-2fba-4599-b6ff-5ee4a4cb32f6" (UID: "360c143b-2fba-4599-b6ff-5ee4a4cb32f6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.232619 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "360c143b-2fba-4599-b6ff-5ee4a4cb32f6" (UID: "360c143b-2fba-4599-b6ff-5ee4a4cb32f6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.235744 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-scripts" (OuterVolumeSpecName: "scripts") pod "360c143b-2fba-4599-b6ff-5ee4a4cb32f6" (UID: "360c143b-2fba-4599-b6ff-5ee4a4cb32f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.240947 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-kube-api-access-48s5z" (OuterVolumeSpecName: "kube-api-access-48s5z") pod "360c143b-2fba-4599-b6ff-5ee4a4cb32f6" (UID: "360c143b-2fba-4599-b6ff-5ee4a4cb32f6"). InnerVolumeSpecName "kube-api-access-48s5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.269578 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.284077 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "360c143b-2fba-4599-b6ff-5ee4a4cb32f6" (UID: "360c143b-2fba-4599-b6ff-5ee4a4cb32f6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.332836 4822 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.332863 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48s5z\" (UniqueName: \"kubernetes.io/projected/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-kube-api-access-48s5z\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.332873 4822 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.332882 4822 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.332890 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.334280 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "360c143b-2fba-4599-b6ff-5ee4a4cb32f6" (UID: "360c143b-2fba-4599-b6ff-5ee4a4cb32f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.378483 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-config-data" (OuterVolumeSpecName: "config-data") pod "360c143b-2fba-4599-b6ff-5ee4a4cb32f6" (UID: "360c143b-2fba-4599-b6ff-5ee4a4cb32f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.433742 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.433772 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/360c143b-2fba-4599-b6ff-5ee4a4cb32f6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.724188 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.734394 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"360c143b-2fba-4599-b6ff-5ee4a4cb32f6","Type":"ContainerDied","Data":"0a0c8da07d8e101b489d2c53645d3abafe8bddc6a32ae7b1442f48aec01c29c7"} Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.734461 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.734486 4822 scope.go:117] "RemoveContainer" containerID="88091be5bc420884d346f15745261462973450a471f0bc4cf4066bf19cb2533e" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.734957 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.814761 4822 scope.go:117] "RemoveContainer" containerID="efe18120e9205b524c22f4cb8078b8a34ff3231d0f7b13943c65d11c0fdf1695" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.814919 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.833875 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.844533 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:22 crc kubenswrapper[4822]: E1124 14:39:22.844928 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="ceilometer-central-agent" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.844946 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="ceilometer-central-agent" Nov 24 14:39:22 crc kubenswrapper[4822]: E1124 14:39:22.844960 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="ceilometer-notification-agent" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.844969 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="ceilometer-notification-agent" Nov 24 14:39:22 crc kubenswrapper[4822]: E1124 14:39:22.844995 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="proxy-httpd" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.845002 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="proxy-httpd" Nov 24 14:39:22 crc kubenswrapper[4822]: E1124 14:39:22.845015 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="sg-core" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.845023 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="sg-core" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.845280 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="proxy-httpd" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.845298 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="sg-core" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.845308 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="ceilometer-central-agent" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.845330 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" containerName="ceilometer-notification-agent" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.848603 4822 scope.go:117] "RemoveContainer" containerID="91dc843e037f676c81fcb7a6d1b291edf12bf9e5f902db6f442d93636a5ff089" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.853301 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.853430 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.860649 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.860821 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:39:22 crc kubenswrapper[4822]: I1124 14:39:22.893362 4822 scope.go:117] "RemoveContainer" containerID="4c5edc72161a570b1de033ee7b676a4e7e5aaaaea501331e1c8a81887e6f64ca" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.045497 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95w8g\" (UniqueName: \"kubernetes.io/projected/9223684f-0e48-41d5-acb4-9ac4d54f9498-kube-api-access-95w8g\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.045544 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-run-httpd\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.045583 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.045646 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-log-httpd\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.045670 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-config-data\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.045685 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.045702 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-scripts\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.147335 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-log-httpd\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.147379 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-config-data\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.147399 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.147420 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-scripts\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.147522 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95w8g\" (UniqueName: \"kubernetes.io/projected/9223684f-0e48-41d5-acb4-9ac4d54f9498-kube-api-access-95w8g\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.147548 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-run-httpd\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.147605 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.147871 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-log-httpd\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.148006 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-run-httpd\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.156047 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.161262 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-config-data\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.166595 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-scripts\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.171839 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.196936 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95w8g\" (UniqueName: \"kubernetes.io/projected/9223684f-0e48-41d5-acb4-9ac4d54f9498-kube-api-access-95w8g\") pod \"ceilometer-0\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.482357 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.725245 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="360c143b-2fba-4599-b6ff-5ee4a4cb32f6" path="/var/lib/kubelet/pods/360c143b-2fba-4599-b6ff-5ee4a4cb32f6/volumes" Nov 24 14:39:23 crc kubenswrapper[4822]: I1124 14:39:23.963414 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:24 crc kubenswrapper[4822]: I1124 14:39:24.030258 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Nov 24 14:39:24 crc kubenswrapper[4822]: I1124 14:39:24.047740 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:24 crc kubenswrapper[4822]: I1124 14:39:24.748440 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerStarted","Data":"02c9857f360b01c7a3e7f0b094e6837bb66279a5535f9d9b4f838f4e3f6bab9e"} Nov 24 14:39:24 crc kubenswrapper[4822]: I1124 14:39:24.748898 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerStarted","Data":"bc0ff4161414c2ce0c5e732d806803ee7efce185bcad63d192cd90c322d07f27"} Nov 24 14:39:24 crc kubenswrapper[4822]: I1124 14:39:24.748482 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:39:24 crc kubenswrapper[4822]: I1124 14:39:24.748930 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:39:24 crc kubenswrapper[4822]: I1124 14:39:24.809008 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:39:25 crc kubenswrapper[4822]: I1124 14:39:25.050706 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:39:25 crc kubenswrapper[4822]: I1124 14:39:25.353608 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:25 crc kubenswrapper[4822]: I1124 14:39:25.353837 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:25 crc kubenswrapper[4822]: I1124 14:39:25.391891 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:25 crc kubenswrapper[4822]: I1124 14:39:25.435752 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:25 crc kubenswrapper[4822]: I1124 14:39:25.759712 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerStarted","Data":"3f8e35a627667f3b8bd6dd1b3c193fee92f85cc46ebe3ab7310ff58bdd69e3f8"} Nov 24 14:39:25 crc kubenswrapper[4822]: I1124 14:39:25.760959 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:25 crc kubenswrapper[4822]: I1124 14:39:25.760998 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:26 crc kubenswrapper[4822]: I1124 14:39:26.769561 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerStarted","Data":"1be4967ea7190c14bc1ec340c848211e256e36230a632c84e10ed1aad75ace7c"} Nov 24 14:39:27 crc kubenswrapper[4822]: I1124 14:39:27.782841 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:39:27 crc kubenswrapper[4822]: I1124 14:39:27.783486 4822 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:39:27 crc kubenswrapper[4822]: I1124 14:39:27.783143 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="ceilometer-central-agent" containerID="cri-o://02c9857f360b01c7a3e7f0b094e6837bb66279a5535f9d9b4f838f4e3f6bab9e" gracePeriod=30 Nov 24 14:39:27 crc kubenswrapper[4822]: I1124 14:39:27.783539 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="sg-core" containerID="cri-o://1be4967ea7190c14bc1ec340c848211e256e36230a632c84e10ed1aad75ace7c" gracePeriod=30 Nov 24 14:39:27 crc kubenswrapper[4822]: I1124 14:39:27.783556 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="proxy-httpd" containerID="cri-o://e05a59adf9f9ecc7a7b1b8349458b639dac6beb704b5489f5ce102767d1338fb" gracePeriod=30 Nov 24 14:39:27 crc kubenswrapper[4822]: I1124 14:39:27.783393 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerStarted","Data":"e05a59adf9f9ecc7a7b1b8349458b639dac6beb704b5489f5ce102767d1338fb"} Nov 24 14:39:27 crc kubenswrapper[4822]: I1124 14:39:27.783650 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="ceilometer-notification-agent" containerID="cri-o://3f8e35a627667f3b8bd6dd1b3c193fee92f85cc46ebe3ab7310ff58bdd69e3f8" gracePeriod=30 Nov 24 14:39:27 crc kubenswrapper[4822]: I1124 14:39:27.783841 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:39:27 crc kubenswrapper[4822]: I1124 14:39:27.816246 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.266430189 podStartE2EDuration="5.816198038s" podCreationTimestamp="2025-11-24 14:39:22 +0000 UTC" firstStartedPulling="2025-11-24 14:39:23.983332107 +0000 UTC m=+1201.099972584" lastFinishedPulling="2025-11-24 14:39:27.533099966 +0000 UTC m=+1204.649740433" observedRunningTime="2025-11-24 14:39:27.800029419 +0000 UTC m=+1204.916669886" watchObservedRunningTime="2025-11-24 14:39:27.816198038 +0000 UTC m=+1204.932838525" Nov 24 14:39:28 crc kubenswrapper[4822]: I1124 14:39:28.036725 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:28 crc kubenswrapper[4822]: I1124 14:39:28.114696 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:39:28 crc kubenswrapper[4822]: I1124 14:39:28.798725 4822 generic.go:334] "Generic (PLEG): container finished" podID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerID="1be4967ea7190c14bc1ec340c848211e256e36230a632c84e10ed1aad75ace7c" exitCode=2 Nov 24 14:39:28 crc kubenswrapper[4822]: I1124 14:39:28.799142 4822 generic.go:334] "Generic (PLEG): container finished" podID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerID="3f8e35a627667f3b8bd6dd1b3c193fee92f85cc46ebe3ab7310ff58bdd69e3f8" exitCode=0 Nov 24 14:39:28 crc kubenswrapper[4822]: I1124 14:39:28.798816 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerDied","Data":"1be4967ea7190c14bc1ec340c848211e256e36230a632c84e10ed1aad75ace7c"} Nov 24 14:39:28 crc kubenswrapper[4822]: I1124 14:39:28.799479 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerDied","Data":"3f8e35a627667f3b8bd6dd1b3c193fee92f85cc46ebe3ab7310ff58bdd69e3f8"} Nov 24 14:39:31 crc kubenswrapper[4822]: I1124 14:39:31.332349 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:39:31 crc kubenswrapper[4822]: E1124 14:39:31.332804 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:39:31 crc kubenswrapper[4822]: E1124 14:39:31.333000 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:39:31 crc kubenswrapper[4822]: E1124 14:39:31.333080 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:40:03.333056053 +0000 UTC m=+1240.449696540 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.186809 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-mr4wn"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.188341 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.216927 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mr4wn"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.262762 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/866995d8-9b1e-4558-b713-955dd3e3621b-operator-scripts\") pod \"nova-api-db-create-mr4wn\" (UID: \"866995d8-9b1e-4558-b713-955dd3e3621b\") " pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.262936 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt9nd\" (UniqueName: \"kubernetes.io/projected/866995d8-9b1e-4558-b713-955dd3e3621b-kube-api-access-qt9nd\") pod \"nova-api-db-create-mr4wn\" (UID: \"866995d8-9b1e-4558-b713-955dd3e3621b\") " pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.364647 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt9nd\" (UniqueName: \"kubernetes.io/projected/866995d8-9b1e-4558-b713-955dd3e3621b-kube-api-access-qt9nd\") pod \"nova-api-db-create-mr4wn\" (UID: \"866995d8-9b1e-4558-b713-955dd3e3621b\") " pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.364763 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/866995d8-9b1e-4558-b713-955dd3e3621b-operator-scripts\") pod \"nova-api-db-create-mr4wn\" (UID: \"866995d8-9b1e-4558-b713-955dd3e3621b\") " pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.365618 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/866995d8-9b1e-4558-b713-955dd3e3621b-operator-scripts\") pod \"nova-api-db-create-mr4wn\" (UID: \"866995d8-9b1e-4558-b713-955dd3e3621b\") " pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.379822 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-nhlp2"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.381128 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.405955 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt9nd\" (UniqueName: \"kubernetes.io/projected/866995d8-9b1e-4558-b713-955dd3e3621b-kube-api-access-qt9nd\") pod \"nova-api-db-create-mr4wn\" (UID: \"866995d8-9b1e-4558-b713-955dd3e3621b\") " pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.407756 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-nhlp2"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.466192 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a102f476-c876-41c0-8700-bac9fbd2b95c-operator-scripts\") pod \"nova-cell0-db-create-nhlp2\" (UID: \"a102f476-c876-41c0-8700-bac9fbd2b95c\") " pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.466277 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qhwj\" (UniqueName: \"kubernetes.io/projected/a102f476-c876-41c0-8700-bac9fbd2b95c-kube-api-access-7qhwj\") pod \"nova-cell0-db-create-nhlp2\" (UID: \"a102f476-c876-41c0-8700-bac9fbd2b95c\") " pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.485121 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c404-account-create-lpw6b"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.486502 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.491559 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.524274 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c404-account-create-lpw6b"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.568899 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxksv\" (UniqueName: \"kubernetes.io/projected/5edd476f-b73c-4f34-b905-baebf09ce800-kube-api-access-lxksv\") pod \"nova-api-c404-account-create-lpw6b\" (UID: \"5edd476f-b73c-4f34-b905-baebf09ce800\") " pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.568975 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5edd476f-b73c-4f34-b905-baebf09ce800-operator-scripts\") pod \"nova-api-c404-account-create-lpw6b\" (UID: \"5edd476f-b73c-4f34-b905-baebf09ce800\") " pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.569097 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a102f476-c876-41c0-8700-bac9fbd2b95c-operator-scripts\") pod \"nova-cell0-db-create-nhlp2\" (UID: \"a102f476-c876-41c0-8700-bac9fbd2b95c\") " pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.569162 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qhwj\" (UniqueName: \"kubernetes.io/projected/a102f476-c876-41c0-8700-bac9fbd2b95c-kube-api-access-7qhwj\") pod \"nova-cell0-db-create-nhlp2\" (UID: \"a102f476-c876-41c0-8700-bac9fbd2b95c\") " pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.570115 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a102f476-c876-41c0-8700-bac9fbd2b95c-operator-scripts\") pod \"nova-cell0-db-create-nhlp2\" (UID: \"a102f476-c876-41c0-8700-bac9fbd2b95c\") " pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.573264 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.589491 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-smgtd"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.590880 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qhwj\" (UniqueName: \"kubernetes.io/projected/a102f476-c876-41c0-8700-bac9fbd2b95c-kube-api-access-7qhwj\") pod \"nova-cell0-db-create-nhlp2\" (UID: \"a102f476-c876-41c0-8700-bac9fbd2b95c\") " pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.590928 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.602163 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-smgtd"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.670557 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5edd476f-b73c-4f34-b905-baebf09ce800-operator-scripts\") pod \"nova-api-c404-account-create-lpw6b\" (UID: \"5edd476f-b73c-4f34-b905-baebf09ce800\") " pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.670982 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxksv\" (UniqueName: \"kubernetes.io/projected/5edd476f-b73c-4f34-b905-baebf09ce800-kube-api-access-lxksv\") pod \"nova-api-c404-account-create-lpw6b\" (UID: \"5edd476f-b73c-4f34-b905-baebf09ce800\") " pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.671901 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5edd476f-b73c-4f34-b905-baebf09ce800-operator-scripts\") pod \"nova-api-c404-account-create-lpw6b\" (UID: \"5edd476f-b73c-4f34-b905-baebf09ce800\") " pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.698441 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxksv\" (UniqueName: \"kubernetes.io/projected/5edd476f-b73c-4f34-b905-baebf09ce800-kube-api-access-lxksv\") pod \"nova-api-c404-account-create-lpw6b\" (UID: \"5edd476f-b73c-4f34-b905-baebf09ce800\") " pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.712287 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-5c5d-account-create-t7dqm"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.715331 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.722017 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.730476 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5c5d-account-create-t7dqm"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.776331 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e7fd483-39ae-40aa-b649-e3558dab68c3-operator-scripts\") pod \"nova-cell0-5c5d-account-create-t7dqm\" (UID: \"6e7fd483-39ae-40aa-b649-e3558dab68c3\") " pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.776474 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86107af1-85fb-4fac-8864-ebdf9d290058-operator-scripts\") pod \"nova-cell1-db-create-smgtd\" (UID: \"86107af1-85fb-4fac-8864-ebdf9d290058\") " pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.776534 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcg7l\" (UniqueName: \"kubernetes.io/projected/86107af1-85fb-4fac-8864-ebdf9d290058-kube-api-access-lcg7l\") pod \"nova-cell1-db-create-smgtd\" (UID: \"86107af1-85fb-4fac-8864-ebdf9d290058\") " pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.776628 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfb5k\" (UniqueName: \"kubernetes.io/projected/6e7fd483-39ae-40aa-b649-e3558dab68c3-kube-api-access-lfb5k\") pod \"nova-cell0-5c5d-account-create-t7dqm\" (UID: \"6e7fd483-39ae-40aa-b649-e3558dab68c3\") " pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.792953 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.812828 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.878930 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e7fd483-39ae-40aa-b649-e3558dab68c3-operator-scripts\") pod \"nova-cell0-5c5d-account-create-t7dqm\" (UID: \"6e7fd483-39ae-40aa-b649-e3558dab68c3\") " pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.878996 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86107af1-85fb-4fac-8864-ebdf9d290058-operator-scripts\") pod \"nova-cell1-db-create-smgtd\" (UID: \"86107af1-85fb-4fac-8864-ebdf9d290058\") " pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.879044 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcg7l\" (UniqueName: \"kubernetes.io/projected/86107af1-85fb-4fac-8864-ebdf9d290058-kube-api-access-lcg7l\") pod \"nova-cell1-db-create-smgtd\" (UID: \"86107af1-85fb-4fac-8864-ebdf9d290058\") " pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.879108 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfb5k\" (UniqueName: \"kubernetes.io/projected/6e7fd483-39ae-40aa-b649-e3558dab68c3-kube-api-access-lfb5k\") pod \"nova-cell0-5c5d-account-create-t7dqm\" (UID: \"6e7fd483-39ae-40aa-b649-e3558dab68c3\") " pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.880553 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86107af1-85fb-4fac-8864-ebdf9d290058-operator-scripts\") pod \"nova-cell1-db-create-smgtd\" (UID: \"86107af1-85fb-4fac-8864-ebdf9d290058\") " pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.880737 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e7fd483-39ae-40aa-b649-e3558dab68c3-operator-scripts\") pod \"nova-cell0-5c5d-account-create-t7dqm\" (UID: \"6e7fd483-39ae-40aa-b649-e3558dab68c3\") " pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.909763 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfb5k\" (UniqueName: \"kubernetes.io/projected/6e7fd483-39ae-40aa-b649-e3558dab68c3-kube-api-access-lfb5k\") pod \"nova-cell0-5c5d-account-create-t7dqm\" (UID: \"6e7fd483-39ae-40aa-b649-e3558dab68c3\") " pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.909841 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-bd5c-account-create-6z47s"] Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.911315 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.913188 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.920433 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcg7l\" (UniqueName: \"kubernetes.io/projected/86107af1-85fb-4fac-8864-ebdf9d290058-kube-api-access-lcg7l\") pod \"nova-cell1-db-create-smgtd\" (UID: \"86107af1-85fb-4fac-8864-ebdf9d290058\") " pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:32 crc kubenswrapper[4822]: I1124 14:39:32.921339 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-bd5c-account-create-6z47s"] Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.054505 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.078680 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.088764 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8frvg\" (UniqueName: \"kubernetes.io/projected/d5ccd4f9-b868-476d-b08f-cc78590964b1-kube-api-access-8frvg\") pod \"nova-cell1-bd5c-account-create-6z47s\" (UID: \"d5ccd4f9-b868-476d-b08f-cc78590964b1\") " pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.089766 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5ccd4f9-b868-476d-b08f-cc78590964b1-operator-scripts\") pod \"nova-cell1-bd5c-account-create-6z47s\" (UID: \"d5ccd4f9-b868-476d-b08f-cc78590964b1\") " pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.105141 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mr4wn"] Nov 24 14:39:33 crc kubenswrapper[4822]: W1124 14:39:33.112352 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod866995d8_9b1e_4558_b713_955dd3e3621b.slice/crio-1701788cb1b1df161c0415f427395e036a020eab1002408d63c5032c3317a6fb WatchSource:0}: Error finding container 1701788cb1b1df161c0415f427395e036a020eab1002408d63c5032c3317a6fb: Status 404 returned error can't find the container with id 1701788cb1b1df161c0415f427395e036a020eab1002408d63c5032c3317a6fb Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.192805 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8frvg\" (UniqueName: \"kubernetes.io/projected/d5ccd4f9-b868-476d-b08f-cc78590964b1-kube-api-access-8frvg\") pod \"nova-cell1-bd5c-account-create-6z47s\" (UID: \"d5ccd4f9-b868-476d-b08f-cc78590964b1\") " pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.192915 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5ccd4f9-b868-476d-b08f-cc78590964b1-operator-scripts\") pod \"nova-cell1-bd5c-account-create-6z47s\" (UID: \"d5ccd4f9-b868-476d-b08f-cc78590964b1\") " pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.193651 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5ccd4f9-b868-476d-b08f-cc78590964b1-operator-scripts\") pod \"nova-cell1-bd5c-account-create-6z47s\" (UID: \"d5ccd4f9-b868-476d-b08f-cc78590964b1\") " pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.213733 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8frvg\" (UniqueName: \"kubernetes.io/projected/d5ccd4f9-b868-476d-b08f-cc78590964b1-kube-api-access-8frvg\") pod \"nova-cell1-bd5c-account-create-6z47s\" (UID: \"d5ccd4f9-b868-476d-b08f-cc78590964b1\") " pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.258244 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.353698 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-nhlp2"] Nov 24 14:39:33 crc kubenswrapper[4822]: W1124 14:39:33.368285 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5edd476f_b73c_4f34_b905_baebf09ce800.slice/crio-d2133756b63156a9ffcefa57d993f93539c172532cd09a1ce243af4074d36cf9 WatchSource:0}: Error finding container d2133756b63156a9ffcefa57d993f93539c172532cd09a1ce243af4074d36cf9: Status 404 returned error can't find the container with id d2133756b63156a9ffcefa57d993f93539c172532cd09a1ce243af4074d36cf9 Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.371398 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c404-account-create-lpw6b"] Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.629885 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-smgtd"] Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.650033 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5c5d-account-create-t7dqm"] Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.876961 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-smgtd" event={"ID":"86107af1-85fb-4fac-8864-ebdf9d290058","Type":"ContainerStarted","Data":"f75b5233475584c55863a180cba7dd32709ff68a718e4e18490a52aae895c08f"} Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.878723 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5c5d-account-create-t7dqm" event={"ID":"6e7fd483-39ae-40aa-b649-e3558dab68c3","Type":"ContainerStarted","Data":"f4d23e422c930a4465dd2689bdb5ca89f2a4a0bb66510e641733a8da0cf8e14a"} Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.883988 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c404-account-create-lpw6b" event={"ID":"5edd476f-b73c-4f34-b905-baebf09ce800","Type":"ContainerStarted","Data":"1fd7145c53cea23eb0f13026d0b0427e31a0dc8964ece07e829785e5a862f6e1"} Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.884051 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c404-account-create-lpw6b" event={"ID":"5edd476f-b73c-4f34-b905-baebf09ce800","Type":"ContainerStarted","Data":"d2133756b63156a9ffcefa57d993f93539c172532cd09a1ce243af4074d36cf9"} Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.887686 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nhlp2" event={"ID":"a102f476-c876-41c0-8700-bac9fbd2b95c","Type":"ContainerStarted","Data":"d526f181dd729f0b676df402b284fef762847ea69c04ff7454d970c587fea0cc"} Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.887715 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nhlp2" event={"ID":"a102f476-c876-41c0-8700-bac9fbd2b95c","Type":"ContainerStarted","Data":"83fd46218e47f5910923eec29082c7372631ba9e08b194ccef68dc31aa8b3dbf"} Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.893911 4822 generic.go:334] "Generic (PLEG): container finished" podID="866995d8-9b1e-4558-b713-955dd3e3621b" containerID="1753807851ea24920b8697e4b5274fdcf13f4f8476bc382c3e60a3279ae0f726" exitCode=0 Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.893963 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mr4wn" event={"ID":"866995d8-9b1e-4558-b713-955dd3e3621b","Type":"ContainerDied","Data":"1753807851ea24920b8697e4b5274fdcf13f4f8476bc382c3e60a3279ae0f726"} Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.893993 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mr4wn" event={"ID":"866995d8-9b1e-4558-b713-955dd3e3621b","Type":"ContainerStarted","Data":"1701788cb1b1df161c0415f427395e036a020eab1002408d63c5032c3317a6fb"} Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.894087 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-bd5c-account-create-6z47s"] Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.897714 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-c404-account-create-lpw6b" podStartSLOduration=1.897691086 podStartE2EDuration="1.897691086s" podCreationTimestamp="2025-11-24 14:39:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:33.896269788 +0000 UTC m=+1211.012910265" watchObservedRunningTime="2025-11-24 14:39:33.897691086 +0000 UTC m=+1211.014331563" Nov 24 14:39:33 crc kubenswrapper[4822]: W1124 14:39:33.903493 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5ccd4f9_b868_476d_b08f_cc78590964b1.slice/crio-c535d0140eb53d349f0c5ce3faf6f439695b1b5bf73c7ba32dffd7b1de30e072 WatchSource:0}: Error finding container c535d0140eb53d349f0c5ce3faf6f439695b1b5bf73c7ba32dffd7b1de30e072: Status 404 returned error can't find the container with id c535d0140eb53d349f0c5ce3faf6f439695b1b5bf73c7ba32dffd7b1de30e072 Nov 24 14:39:33 crc kubenswrapper[4822]: I1124 14:39:33.913839 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-nhlp2" podStartSLOduration=1.913817314 podStartE2EDuration="1.913817314s" podCreationTimestamp="2025-11-24 14:39:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:33.911671746 +0000 UTC m=+1211.028312223" watchObservedRunningTime="2025-11-24 14:39:33.913817314 +0000 UTC m=+1211.030457791" Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.903660 4822 generic.go:334] "Generic (PLEG): container finished" podID="6e7fd483-39ae-40aa-b649-e3558dab68c3" containerID="e4b885cd4ea462207162e2986589f5ca3e09df63d17cec2252b6d012edec94f5" exitCode=0 Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.903761 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5c5d-account-create-t7dqm" event={"ID":"6e7fd483-39ae-40aa-b649-e3558dab68c3","Type":"ContainerDied","Data":"e4b885cd4ea462207162e2986589f5ca3e09df63d17cec2252b6d012edec94f5"} Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.906219 4822 generic.go:334] "Generic (PLEG): container finished" podID="5edd476f-b73c-4f34-b905-baebf09ce800" containerID="1fd7145c53cea23eb0f13026d0b0427e31a0dc8964ece07e829785e5a862f6e1" exitCode=0 Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.906252 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c404-account-create-lpw6b" event={"ID":"5edd476f-b73c-4f34-b905-baebf09ce800","Type":"ContainerDied","Data":"1fd7145c53cea23eb0f13026d0b0427e31a0dc8964ece07e829785e5a862f6e1"} Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.907833 4822 generic.go:334] "Generic (PLEG): container finished" podID="a102f476-c876-41c0-8700-bac9fbd2b95c" containerID="d526f181dd729f0b676df402b284fef762847ea69c04ff7454d970c587fea0cc" exitCode=0 Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.907896 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nhlp2" event={"ID":"a102f476-c876-41c0-8700-bac9fbd2b95c","Type":"ContainerDied","Data":"d526f181dd729f0b676df402b284fef762847ea69c04ff7454d970c587fea0cc"} Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.909422 4822 generic.go:334] "Generic (PLEG): container finished" podID="d5ccd4f9-b868-476d-b08f-cc78590964b1" containerID="07d9bdb91d4ba52bb4e67aed971b71b5a2f155b692bcdef13a2832a4efafdbce" exitCode=0 Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.909505 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-bd5c-account-create-6z47s" event={"ID":"d5ccd4f9-b868-476d-b08f-cc78590964b1","Type":"ContainerDied","Data":"07d9bdb91d4ba52bb4e67aed971b71b5a2f155b692bcdef13a2832a4efafdbce"} Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.909544 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-bd5c-account-create-6z47s" event={"ID":"d5ccd4f9-b868-476d-b08f-cc78590964b1","Type":"ContainerStarted","Data":"c535d0140eb53d349f0c5ce3faf6f439695b1b5bf73c7ba32dffd7b1de30e072"} Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.910770 4822 generic.go:334] "Generic (PLEG): container finished" podID="86107af1-85fb-4fac-8864-ebdf9d290058" containerID="71994cbbfdd0df3867c8c71fd5f7c9461f2e2eb0d7357f6143be0b0c4fc71675" exitCode=0 Nov 24 14:39:34 crc kubenswrapper[4822]: I1124 14:39:34.910799 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-smgtd" event={"ID":"86107af1-85fb-4fac-8864-ebdf9d290058","Type":"ContainerDied","Data":"71994cbbfdd0df3867c8c71fd5f7c9461f2e2eb0d7357f6143be0b0c4fc71675"} Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.289056 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.354563 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt9nd\" (UniqueName: \"kubernetes.io/projected/866995d8-9b1e-4558-b713-955dd3e3621b-kube-api-access-qt9nd\") pod \"866995d8-9b1e-4558-b713-955dd3e3621b\" (UID: \"866995d8-9b1e-4558-b713-955dd3e3621b\") " Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.354930 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/866995d8-9b1e-4558-b713-955dd3e3621b-operator-scripts\") pod \"866995d8-9b1e-4558-b713-955dd3e3621b\" (UID: \"866995d8-9b1e-4558-b713-955dd3e3621b\") " Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.355431 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/866995d8-9b1e-4558-b713-955dd3e3621b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "866995d8-9b1e-4558-b713-955dd3e3621b" (UID: "866995d8-9b1e-4558-b713-955dd3e3621b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.355567 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/866995d8-9b1e-4558-b713-955dd3e3621b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.361456 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/866995d8-9b1e-4558-b713-955dd3e3621b-kube-api-access-qt9nd" (OuterVolumeSpecName: "kube-api-access-qt9nd") pod "866995d8-9b1e-4558-b713-955dd3e3621b" (UID: "866995d8-9b1e-4558-b713-955dd3e3621b"). InnerVolumeSpecName "kube-api-access-qt9nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.457099 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt9nd\" (UniqueName: \"kubernetes.io/projected/866995d8-9b1e-4558-b713-955dd3e3621b-kube-api-access-qt9nd\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.923615 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mr4wn" event={"ID":"866995d8-9b1e-4558-b713-955dd3e3621b","Type":"ContainerDied","Data":"1701788cb1b1df161c0415f427395e036a020eab1002408d63c5032c3317a6fb"} Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.923671 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1701788cb1b1df161c0415f427395e036a020eab1002408d63c5032c3317a6fb" Nov 24 14:39:35 crc kubenswrapper[4822]: I1124 14:39:35.923865 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mr4wn" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.312833 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.480743 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxksv\" (UniqueName: \"kubernetes.io/projected/5edd476f-b73c-4f34-b905-baebf09ce800-kube-api-access-lxksv\") pod \"5edd476f-b73c-4f34-b905-baebf09ce800\" (UID: \"5edd476f-b73c-4f34-b905-baebf09ce800\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.480862 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5edd476f-b73c-4f34-b905-baebf09ce800-operator-scripts\") pod \"5edd476f-b73c-4f34-b905-baebf09ce800\" (UID: \"5edd476f-b73c-4f34-b905-baebf09ce800\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.481945 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5edd476f-b73c-4f34-b905-baebf09ce800-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5edd476f-b73c-4f34-b905-baebf09ce800" (UID: "5edd476f-b73c-4f34-b905-baebf09ce800"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.488424 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5edd476f-b73c-4f34-b905-baebf09ce800-kube-api-access-lxksv" (OuterVolumeSpecName: "kube-api-access-lxksv") pod "5edd476f-b73c-4f34-b905-baebf09ce800" (UID: "5edd476f-b73c-4f34-b905-baebf09ce800"). InnerVolumeSpecName "kube-api-access-lxksv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.589859 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxksv\" (UniqueName: \"kubernetes.io/projected/5edd476f-b73c-4f34-b905-baebf09ce800-kube-api-access-lxksv\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.589889 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5edd476f-b73c-4f34-b905-baebf09ce800-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.644847 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.651772 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.660027 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.671772 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.690669 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5ccd4f9-b868-476d-b08f-cc78590964b1-operator-scripts\") pod \"d5ccd4f9-b868-476d-b08f-cc78590964b1\" (UID: \"d5ccd4f9-b868-476d-b08f-cc78590964b1\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.690769 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8frvg\" (UniqueName: \"kubernetes.io/projected/d5ccd4f9-b868-476d-b08f-cc78590964b1-kube-api-access-8frvg\") pod \"d5ccd4f9-b868-476d-b08f-cc78590964b1\" (UID: \"d5ccd4f9-b868-476d-b08f-cc78590964b1\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.690812 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfb5k\" (UniqueName: \"kubernetes.io/projected/6e7fd483-39ae-40aa-b649-e3558dab68c3-kube-api-access-lfb5k\") pod \"6e7fd483-39ae-40aa-b649-e3558dab68c3\" (UID: \"6e7fd483-39ae-40aa-b649-e3558dab68c3\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.690854 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e7fd483-39ae-40aa-b649-e3558dab68c3-operator-scripts\") pod \"6e7fd483-39ae-40aa-b649-e3558dab68c3\" (UID: \"6e7fd483-39ae-40aa-b649-e3558dab68c3\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.690962 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86107af1-85fb-4fac-8864-ebdf9d290058-operator-scripts\") pod \"86107af1-85fb-4fac-8864-ebdf9d290058\" (UID: \"86107af1-85fb-4fac-8864-ebdf9d290058\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.691041 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qhwj\" (UniqueName: \"kubernetes.io/projected/a102f476-c876-41c0-8700-bac9fbd2b95c-kube-api-access-7qhwj\") pod \"a102f476-c876-41c0-8700-bac9fbd2b95c\" (UID: \"a102f476-c876-41c0-8700-bac9fbd2b95c\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.691085 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a102f476-c876-41c0-8700-bac9fbd2b95c-operator-scripts\") pod \"a102f476-c876-41c0-8700-bac9fbd2b95c\" (UID: \"a102f476-c876-41c0-8700-bac9fbd2b95c\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.691163 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcg7l\" (UniqueName: \"kubernetes.io/projected/86107af1-85fb-4fac-8864-ebdf9d290058-kube-api-access-lcg7l\") pod \"86107af1-85fb-4fac-8864-ebdf9d290058\" (UID: \"86107af1-85fb-4fac-8864-ebdf9d290058\") " Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.691168 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5ccd4f9-b868-476d-b08f-cc78590964b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d5ccd4f9-b868-476d-b08f-cc78590964b1" (UID: "d5ccd4f9-b868-476d-b08f-cc78590964b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.691444 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e7fd483-39ae-40aa-b649-e3558dab68c3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6e7fd483-39ae-40aa-b649-e3558dab68c3" (UID: "6e7fd483-39ae-40aa-b649-e3558dab68c3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.691768 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5ccd4f9-b868-476d-b08f-cc78590964b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.691787 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e7fd483-39ae-40aa-b649-e3558dab68c3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.692056 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a102f476-c876-41c0-8700-bac9fbd2b95c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a102f476-c876-41c0-8700-bac9fbd2b95c" (UID: "a102f476-c876-41c0-8700-bac9fbd2b95c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.692082 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86107af1-85fb-4fac-8864-ebdf9d290058-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "86107af1-85fb-4fac-8864-ebdf9d290058" (UID: "86107af1-85fb-4fac-8864-ebdf9d290058"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.694960 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e7fd483-39ae-40aa-b649-e3558dab68c3-kube-api-access-lfb5k" (OuterVolumeSpecName: "kube-api-access-lfb5k") pod "6e7fd483-39ae-40aa-b649-e3558dab68c3" (UID: "6e7fd483-39ae-40aa-b649-e3558dab68c3"). InnerVolumeSpecName "kube-api-access-lfb5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.696367 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86107af1-85fb-4fac-8864-ebdf9d290058-kube-api-access-lcg7l" (OuterVolumeSpecName: "kube-api-access-lcg7l") pod "86107af1-85fb-4fac-8864-ebdf9d290058" (UID: "86107af1-85fb-4fac-8864-ebdf9d290058"). InnerVolumeSpecName "kube-api-access-lcg7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.696537 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a102f476-c876-41c0-8700-bac9fbd2b95c-kube-api-access-7qhwj" (OuterVolumeSpecName: "kube-api-access-7qhwj") pod "a102f476-c876-41c0-8700-bac9fbd2b95c" (UID: "a102f476-c876-41c0-8700-bac9fbd2b95c"). InnerVolumeSpecName "kube-api-access-7qhwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.696617 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5ccd4f9-b868-476d-b08f-cc78590964b1-kube-api-access-8frvg" (OuterVolumeSpecName: "kube-api-access-8frvg") pod "d5ccd4f9-b868-476d-b08f-cc78590964b1" (UID: "d5ccd4f9-b868-476d-b08f-cc78590964b1"). InnerVolumeSpecName "kube-api-access-8frvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.793141 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcg7l\" (UniqueName: \"kubernetes.io/projected/86107af1-85fb-4fac-8864-ebdf9d290058-kube-api-access-lcg7l\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.793402 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8frvg\" (UniqueName: \"kubernetes.io/projected/d5ccd4f9-b868-476d-b08f-cc78590964b1-kube-api-access-8frvg\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.793412 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfb5k\" (UniqueName: \"kubernetes.io/projected/6e7fd483-39ae-40aa-b649-e3558dab68c3-kube-api-access-lfb5k\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.793423 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86107af1-85fb-4fac-8864-ebdf9d290058-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.793432 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qhwj\" (UniqueName: \"kubernetes.io/projected/a102f476-c876-41c0-8700-bac9fbd2b95c-kube-api-access-7qhwj\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.793441 4822 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a102f476-c876-41c0-8700-bac9fbd2b95c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.936542 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c404-account-create-lpw6b" event={"ID":"5edd476f-b73c-4f34-b905-baebf09ce800","Type":"ContainerDied","Data":"d2133756b63156a9ffcefa57d993f93539c172532cd09a1ce243af4074d36cf9"} Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.937751 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2133756b63156a9ffcefa57d993f93539c172532cd09a1ce243af4074d36cf9" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.937910 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c404-account-create-lpw6b" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.939618 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nhlp2" event={"ID":"a102f476-c876-41c0-8700-bac9fbd2b95c","Type":"ContainerDied","Data":"83fd46218e47f5910923eec29082c7372631ba9e08b194ccef68dc31aa8b3dbf"} Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.939667 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83fd46218e47f5910923eec29082c7372631ba9e08b194ccef68dc31aa8b3dbf" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.939732 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nhlp2" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.943716 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-bd5c-account-create-6z47s" event={"ID":"d5ccd4f9-b868-476d-b08f-cc78590964b1","Type":"ContainerDied","Data":"c535d0140eb53d349f0c5ce3faf6f439695b1b5bf73c7ba32dffd7b1de30e072"} Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.943742 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c535d0140eb53d349f0c5ce3faf6f439695b1b5bf73c7ba32dffd7b1de30e072" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.943794 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-bd5c-account-create-6z47s" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.949272 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-smgtd" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.949284 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-smgtd" event={"ID":"86107af1-85fb-4fac-8864-ebdf9d290058","Type":"ContainerDied","Data":"f75b5233475584c55863a180cba7dd32709ff68a718e4e18490a52aae895c08f"} Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.949356 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f75b5233475584c55863a180cba7dd32709ff68a718e4e18490a52aae895c08f" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.957780 4822 generic.go:334] "Generic (PLEG): container finished" podID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerID="02c9857f360b01c7a3e7f0b094e6837bb66279a5535f9d9b4f838f4e3f6bab9e" exitCode=0 Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.957862 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerDied","Data":"02c9857f360b01c7a3e7f0b094e6837bb66279a5535f9d9b4f838f4e3f6bab9e"} Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.962017 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5c5d-account-create-t7dqm" event={"ID":"6e7fd483-39ae-40aa-b649-e3558dab68c3","Type":"ContainerDied","Data":"f4d23e422c930a4465dd2689bdb5ca89f2a4a0bb66510e641733a8da0cf8e14a"} Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.962049 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4d23e422c930a4465dd2689bdb5ca89f2a4a0bb66510e641733a8da0cf8e14a" Nov 24 14:39:36 crc kubenswrapper[4822]: I1124 14:39:36.962118 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5c5d-account-create-t7dqm" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.950678 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7kfv9"] Nov 24 14:39:37 crc kubenswrapper[4822]: E1124 14:39:37.951097 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e7fd483-39ae-40aa-b649-e3558dab68c3" containerName="mariadb-account-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951110 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e7fd483-39ae-40aa-b649-e3558dab68c3" containerName="mariadb-account-create" Nov 24 14:39:37 crc kubenswrapper[4822]: E1124 14:39:37.951131 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86107af1-85fb-4fac-8864-ebdf9d290058" containerName="mariadb-database-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951137 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="86107af1-85fb-4fac-8864-ebdf9d290058" containerName="mariadb-database-create" Nov 24 14:39:37 crc kubenswrapper[4822]: E1124 14:39:37.951154 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5edd476f-b73c-4f34-b905-baebf09ce800" containerName="mariadb-account-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951160 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5edd476f-b73c-4f34-b905-baebf09ce800" containerName="mariadb-account-create" Nov 24 14:39:37 crc kubenswrapper[4822]: E1124 14:39:37.951172 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866995d8-9b1e-4558-b713-955dd3e3621b" containerName="mariadb-database-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951178 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="866995d8-9b1e-4558-b713-955dd3e3621b" containerName="mariadb-database-create" Nov 24 14:39:37 crc kubenswrapper[4822]: E1124 14:39:37.951193 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5ccd4f9-b868-476d-b08f-cc78590964b1" containerName="mariadb-account-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951198 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5ccd4f9-b868-476d-b08f-cc78590964b1" containerName="mariadb-account-create" Nov 24 14:39:37 crc kubenswrapper[4822]: E1124 14:39:37.951225 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a102f476-c876-41c0-8700-bac9fbd2b95c" containerName="mariadb-database-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951231 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a102f476-c876-41c0-8700-bac9fbd2b95c" containerName="mariadb-database-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951410 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="5edd476f-b73c-4f34-b905-baebf09ce800" containerName="mariadb-account-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951428 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="86107af1-85fb-4fac-8864-ebdf9d290058" containerName="mariadb-database-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951440 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5ccd4f9-b868-476d-b08f-cc78590964b1" containerName="mariadb-account-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951450 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e7fd483-39ae-40aa-b649-e3558dab68c3" containerName="mariadb-account-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951460 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a102f476-c876-41c0-8700-bac9fbd2b95c" containerName="mariadb-database-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.951472 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="866995d8-9b1e-4558-b713-955dd3e3621b" containerName="mariadb-database-create" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.952243 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.953620 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-mjmnb" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.954842 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.960048 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 14:39:37 crc kubenswrapper[4822]: I1124 14:39:37.960549 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7kfv9"] Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.019729 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-scripts\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.019903 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.020064 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-config-data\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.020100 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbrff\" (UniqueName: \"kubernetes.io/projected/8c89b401-31c4-46b1-aa12-49e823e2cf21-kube-api-access-bbrff\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.122794 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.123151 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-config-data\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.123173 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbrff\" (UniqueName: \"kubernetes.io/projected/8c89b401-31c4-46b1-aa12-49e823e2cf21-kube-api-access-bbrff\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.123259 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-scripts\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.129090 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-config-data\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.129099 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-scripts\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.137369 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.138309 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbrff\" (UniqueName: \"kubernetes.io/projected/8c89b401-31c4-46b1-aa12-49e823e2cf21-kube-api-access-bbrff\") pod \"nova-cell0-conductor-db-sync-7kfv9\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.268849 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.839924 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7kfv9"] Nov 24 14:39:38 crc kubenswrapper[4822]: I1124 14:39:38.980001 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7kfv9" event={"ID":"8c89b401-31c4-46b1-aa12-49e823e2cf21","Type":"ContainerStarted","Data":"c01bf376849a923303f2ea3c4a9ffa0068206164f97f99b3bf0f65bb6593ab8f"} Nov 24 14:39:48 crc kubenswrapper[4822]: I1124 14:39:48.085959 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7kfv9" event={"ID":"8c89b401-31c4-46b1-aa12-49e823e2cf21","Type":"ContainerStarted","Data":"2f267ad20a6e8b56fd209664fa21f4c243b40fba5cfae1eea9b88b8d08ffd871"} Nov 24 14:39:48 crc kubenswrapper[4822]: I1124 14:39:48.122270 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7kfv9" podStartSLOduration=3.15907284 podStartE2EDuration="11.122243117s" podCreationTimestamp="2025-11-24 14:39:37 +0000 UTC" firstStartedPulling="2025-11-24 14:39:38.840816783 +0000 UTC m=+1215.957457260" lastFinishedPulling="2025-11-24 14:39:46.80398706 +0000 UTC m=+1223.920627537" observedRunningTime="2025-11-24 14:39:48.11081147 +0000 UTC m=+1225.227451987" watchObservedRunningTime="2025-11-24 14:39:48.122243117 +0000 UTC m=+1225.238883604" Nov 24 14:39:53 crc kubenswrapper[4822]: I1124 14:39:53.489181 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 14:39:57 crc kubenswrapper[4822]: I1124 14:39:57.193054 4822 generic.go:334] "Generic (PLEG): container finished" podID="8c89b401-31c4-46b1-aa12-49e823e2cf21" containerID="2f267ad20a6e8b56fd209664fa21f4c243b40fba5cfae1eea9b88b8d08ffd871" exitCode=0 Nov 24 14:39:57 crc kubenswrapper[4822]: I1124 14:39:57.193133 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7kfv9" event={"ID":"8c89b401-31c4-46b1-aa12-49e823e2cf21","Type":"ContainerDied","Data":"2f267ad20a6e8b56fd209664fa21f4c243b40fba5cfae1eea9b88b8d08ffd871"} Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.208545 4822 generic.go:334] "Generic (PLEG): container finished" podID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerID="e05a59adf9f9ecc7a7b1b8349458b639dac6beb704b5489f5ce102767d1338fb" exitCode=137 Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.208982 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerDied","Data":"e05a59adf9f9ecc7a7b1b8349458b639dac6beb704b5489f5ce102767d1338fb"} Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.425654 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.604697 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-scripts\") pod \"9223684f-0e48-41d5-acb4-9ac4d54f9498\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.604799 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-log-httpd\") pod \"9223684f-0e48-41d5-acb4-9ac4d54f9498\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.604860 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95w8g\" (UniqueName: \"kubernetes.io/projected/9223684f-0e48-41d5-acb4-9ac4d54f9498-kube-api-access-95w8g\") pod \"9223684f-0e48-41d5-acb4-9ac4d54f9498\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.604894 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-run-httpd\") pod \"9223684f-0e48-41d5-acb4-9ac4d54f9498\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.604981 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-combined-ca-bundle\") pod \"9223684f-0e48-41d5-acb4-9ac4d54f9498\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.605016 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-config-data\") pod \"9223684f-0e48-41d5-acb4-9ac4d54f9498\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.605041 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-sg-core-conf-yaml\") pod \"9223684f-0e48-41d5-acb4-9ac4d54f9498\" (UID: \"9223684f-0e48-41d5-acb4-9ac4d54f9498\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.606420 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9223684f-0e48-41d5-acb4-9ac4d54f9498" (UID: "9223684f-0e48-41d5-acb4-9ac4d54f9498"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.607496 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9223684f-0e48-41d5-acb4-9ac4d54f9498" (UID: "9223684f-0e48-41d5-acb4-9ac4d54f9498"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.623511 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-scripts" (OuterVolumeSpecName: "scripts") pod "9223684f-0e48-41d5-acb4-9ac4d54f9498" (UID: "9223684f-0e48-41d5-acb4-9ac4d54f9498"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.623697 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9223684f-0e48-41d5-acb4-9ac4d54f9498-kube-api-access-95w8g" (OuterVolumeSpecName: "kube-api-access-95w8g") pod "9223684f-0e48-41d5-acb4-9ac4d54f9498" (UID: "9223684f-0e48-41d5-acb4-9ac4d54f9498"). InnerVolumeSpecName "kube-api-access-95w8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.650286 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9223684f-0e48-41d5-acb4-9ac4d54f9498" (UID: "9223684f-0e48-41d5-acb4-9ac4d54f9498"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.655168 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.705241 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9223684f-0e48-41d5-acb4-9ac4d54f9498" (UID: "9223684f-0e48-41d5-acb4-9ac4d54f9498"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.708102 4822 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.708143 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95w8g\" (UniqueName: \"kubernetes.io/projected/9223684f-0e48-41d5-acb4-9ac4d54f9498-kube-api-access-95w8g\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.708159 4822 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9223684f-0e48-41d5-acb4-9ac4d54f9498-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.708173 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.708186 4822 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.708198 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.725294 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-config-data" (OuterVolumeSpecName: "config-data") pod "9223684f-0e48-41d5-acb4-9ac4d54f9498" (UID: "9223684f-0e48-41d5-acb4-9ac4d54f9498"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.809512 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-config-data\") pod \"8c89b401-31c4-46b1-aa12-49e823e2cf21\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.809668 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-scripts\") pod \"8c89b401-31c4-46b1-aa12-49e823e2cf21\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.809728 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbrff\" (UniqueName: \"kubernetes.io/projected/8c89b401-31c4-46b1-aa12-49e823e2cf21-kube-api-access-bbrff\") pod \"8c89b401-31c4-46b1-aa12-49e823e2cf21\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.809885 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-combined-ca-bundle\") pod \"8c89b401-31c4-46b1-aa12-49e823e2cf21\" (UID: \"8c89b401-31c4-46b1-aa12-49e823e2cf21\") " Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.810493 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223684f-0e48-41d5-acb4-9ac4d54f9498-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.813818 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-scripts" (OuterVolumeSpecName: "scripts") pod "8c89b401-31c4-46b1-aa12-49e823e2cf21" (UID: "8c89b401-31c4-46b1-aa12-49e823e2cf21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.813905 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c89b401-31c4-46b1-aa12-49e823e2cf21-kube-api-access-bbrff" (OuterVolumeSpecName: "kube-api-access-bbrff") pod "8c89b401-31c4-46b1-aa12-49e823e2cf21" (UID: "8c89b401-31c4-46b1-aa12-49e823e2cf21"). InnerVolumeSpecName "kube-api-access-bbrff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.840198 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c89b401-31c4-46b1-aa12-49e823e2cf21" (UID: "8c89b401-31c4-46b1-aa12-49e823e2cf21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.851372 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-config-data" (OuterVolumeSpecName: "config-data") pod "8c89b401-31c4-46b1-aa12-49e823e2cf21" (UID: "8c89b401-31c4-46b1-aa12-49e823e2cf21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.913008 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.913053 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.913067 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c89b401-31c4-46b1-aa12-49e823e2cf21-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:58 crc kubenswrapper[4822]: I1124 14:39:58.913078 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbrff\" (UniqueName: \"kubernetes.io/projected/8c89b401-31c4-46b1-aa12-49e823e2cf21-kube-api-access-bbrff\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.224581 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9223684f-0e48-41d5-acb4-9ac4d54f9498","Type":"ContainerDied","Data":"bc0ff4161414c2ce0c5e732d806803ee7efce185bcad63d192cd90c322d07f27"} Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.224658 4822 scope.go:117] "RemoveContainer" containerID="e05a59adf9f9ecc7a7b1b8349458b639dac6beb704b5489f5ce102767d1338fb" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.224862 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.239935 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7kfv9" event={"ID":"8c89b401-31c4-46b1-aa12-49e823e2cf21","Type":"ContainerDied","Data":"c01bf376849a923303f2ea3c4a9ffa0068206164f97f99b3bf0f65bb6593ab8f"} Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.240239 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c01bf376849a923303f2ea3c4a9ffa0068206164f97f99b3bf0f65bb6593ab8f" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.240423 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7kfv9" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.271674 4822 scope.go:117] "RemoveContainer" containerID="1be4967ea7190c14bc1ec340c848211e256e36230a632c84e10ed1aad75ace7c" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.311094 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.315075 4822 scope.go:117] "RemoveContainer" containerID="3f8e35a627667f3b8bd6dd1b3c193fee92f85cc46ebe3ab7310ff58bdd69e3f8" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.354288 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.367297 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:59 crc kubenswrapper[4822]: E1124 14:39:59.367986 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="ceilometer-notification-agent" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368004 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="ceilometer-notification-agent" Nov 24 14:39:59 crc kubenswrapper[4822]: E1124 14:39:59.368021 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="ceilometer-central-agent" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368028 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="ceilometer-central-agent" Nov 24 14:39:59 crc kubenswrapper[4822]: E1124 14:39:59.368042 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="proxy-httpd" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368050 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="proxy-httpd" Nov 24 14:39:59 crc kubenswrapper[4822]: E1124 14:39:59.368068 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c89b401-31c4-46b1-aa12-49e823e2cf21" containerName="nova-cell0-conductor-db-sync" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368074 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c89b401-31c4-46b1-aa12-49e823e2cf21" containerName="nova-cell0-conductor-db-sync" Nov 24 14:39:59 crc kubenswrapper[4822]: E1124 14:39:59.368087 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="sg-core" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368092 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="sg-core" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368299 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="ceilometer-central-agent" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368316 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="sg-core" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368324 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c89b401-31c4-46b1-aa12-49e823e2cf21" containerName="nova-cell0-conductor-db-sync" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368340 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="proxy-httpd" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.368349 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" containerName="ceilometer-notification-agent" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.370110 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.370687 4822 scope.go:117] "RemoveContainer" containerID="02c9857f360b01c7a3e7f0b094e6837bb66279a5535f9d9b4f838f4e3f6bab9e" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.387441 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.388248 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.396617 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.411471 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.413844 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.419603 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.419850 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-mjmnb" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.439946 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.525972 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec086ba-ec9d-4267-a5e4-0beac4c89184-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3ec086ba-ec9d-4267-a5e4-0beac4c89184\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.526031 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-log-httpd\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.526054 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qflwj\" (UniqueName: \"kubernetes.io/projected/3ec086ba-ec9d-4267-a5e4-0beac4c89184-kube-api-access-qflwj\") pod \"nova-cell0-conductor-0\" (UID: \"3ec086ba-ec9d-4267-a5e4-0beac4c89184\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.526076 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-scripts\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.526347 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqh98\" (UniqueName: \"kubernetes.io/projected/31d97142-bc40-4398-b4c1-fe12a14f7398-kube-api-access-cqh98\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.526422 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-run-httpd\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.526689 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.526751 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.526835 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-config-data\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.527103 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ec086ba-ec9d-4267-a5e4-0beac4c89184-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3ec086ba-ec9d-4267-a5e4-0beac4c89184\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.628733 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.628798 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.628834 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-config-data\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.628908 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ec086ba-ec9d-4267-a5e4-0beac4c89184-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3ec086ba-ec9d-4267-a5e4-0beac4c89184\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.628932 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec086ba-ec9d-4267-a5e4-0beac4c89184-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3ec086ba-ec9d-4267-a5e4-0beac4c89184\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.628958 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-log-httpd\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.628976 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qflwj\" (UniqueName: \"kubernetes.io/projected/3ec086ba-ec9d-4267-a5e4-0beac4c89184-kube-api-access-qflwj\") pod \"nova-cell0-conductor-0\" (UID: \"3ec086ba-ec9d-4267-a5e4-0beac4c89184\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.628997 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-scripts\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.629027 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqh98\" (UniqueName: \"kubernetes.io/projected/31d97142-bc40-4398-b4c1-fe12a14f7398-kube-api-access-cqh98\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.629047 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-run-httpd\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.629535 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-run-httpd\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.631126 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-log-httpd\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.635669 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.635748 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-scripts\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.635673 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.636742 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-config-data\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.645918 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec086ba-ec9d-4267-a5e4-0beac4c89184-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3ec086ba-ec9d-4267-a5e4-0beac4c89184\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.647020 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qflwj\" (UniqueName: \"kubernetes.io/projected/3ec086ba-ec9d-4267-a5e4-0beac4c89184-kube-api-access-qflwj\") pod \"nova-cell0-conductor-0\" (UID: \"3ec086ba-ec9d-4267-a5e4-0beac4c89184\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.648183 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ec086ba-ec9d-4267-a5e4-0beac4c89184-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3ec086ba-ec9d-4267-a5e4-0beac4c89184\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.650403 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqh98\" (UniqueName: \"kubernetes.io/projected/31d97142-bc40-4398-b4c1-fe12a14f7398-kube-api-access-cqh98\") pod \"ceilometer-0\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.710153 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.723826 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9223684f-0e48-41d5-acb4-9ac4d54f9498" path="/var/lib/kubelet/pods/9223684f-0e48-41d5-acb4-9ac4d54f9498/volumes" Nov 24 14:39:59 crc kubenswrapper[4822]: I1124 14:39:59.747450 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:40:00 crc kubenswrapper[4822]: I1124 14:40:00.275238 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:00 crc kubenswrapper[4822]: I1124 14:40:00.290647 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:40:01 crc kubenswrapper[4822]: I1124 14:40:01.266675 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3ec086ba-ec9d-4267-a5e4-0beac4c89184","Type":"ContainerStarted","Data":"a96d1089a7e057b8d899843f3bc054581e7bb7c49aaf32dd4281aa96a4c020f7"} Nov 24 14:40:01 crc kubenswrapper[4822]: I1124 14:40:01.267322 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 14:40:01 crc kubenswrapper[4822]: I1124 14:40:01.267339 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3ec086ba-ec9d-4267-a5e4-0beac4c89184","Type":"ContainerStarted","Data":"a0c23f2f0724590666c480e626fee728c4b1255982a004452bc7194c9761b1d9"} Nov 24 14:40:01 crc kubenswrapper[4822]: I1124 14:40:01.270098 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerStarted","Data":"802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037"} Nov 24 14:40:01 crc kubenswrapper[4822]: I1124 14:40:01.270154 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerStarted","Data":"d1cbb61f4f161385b19ceee63778d5d313495e32301479c0e70ce98d18f13fb0"} Nov 24 14:40:01 crc kubenswrapper[4822]: I1124 14:40:01.301318 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.301292714 podStartE2EDuration="2.301292714s" podCreationTimestamp="2025-11-24 14:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:01.291641358 +0000 UTC m=+1238.408281835" watchObservedRunningTime="2025-11-24 14:40:01.301292714 +0000 UTC m=+1238.417933191" Nov 24 14:40:02 crc kubenswrapper[4822]: I1124 14:40:02.283910 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerStarted","Data":"9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80"} Nov 24 14:40:03 crc kubenswrapper[4822]: I1124 14:40:03.294491 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerStarted","Data":"770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90"} Nov 24 14:40:03 crc kubenswrapper[4822]: I1124 14:40:03.411716 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:40:03 crc kubenswrapper[4822]: E1124 14:40:03.411910 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:40:03 crc kubenswrapper[4822]: E1124 14:40:03.411934 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:40:03 crc kubenswrapper[4822]: E1124 14:40:03.412006 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:41:07.411986018 +0000 UTC m=+1304.528626485 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:40:04 crc kubenswrapper[4822]: I1124 14:40:04.307871 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerStarted","Data":"cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884"} Nov 24 14:40:04 crc kubenswrapper[4822]: I1124 14:40:04.308388 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:40:04 crc kubenswrapper[4822]: I1124 14:40:04.333552 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8360653789999999 podStartE2EDuration="5.333533834s" podCreationTimestamp="2025-11-24 14:39:59 +0000 UTC" firstStartedPulling="2025-11-24 14:40:00.276710481 +0000 UTC m=+1237.393350988" lastFinishedPulling="2025-11-24 14:40:03.774178926 +0000 UTC m=+1240.890819443" observedRunningTime="2025-11-24 14:40:04.327247055 +0000 UTC m=+1241.443887622" watchObservedRunningTime="2025-11-24 14:40:04.333533834 +0000 UTC m=+1241.450174311" Nov 24 14:40:09 crc kubenswrapper[4822]: I1124 14:40:09.881758 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.423872 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-8p47j"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.425474 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.430046 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.430368 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.437732 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8p47j"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.492467 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.492547 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-config-data\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.492747 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk9d2\" (UniqueName: \"kubernetes.io/projected/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-kube-api-access-gk9d2\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.492847 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-scripts\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.563812 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.565124 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.566512 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.576458 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.594508 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.594579 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sct4t\" (UniqueName: \"kubernetes.io/projected/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-kube-api-access-sct4t\") pod \"nova-scheduler-0\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.594613 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-config-data\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.594657 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk9d2\" (UniqueName: \"kubernetes.io/projected/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-kube-api-access-gk9d2\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.594692 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-scripts\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.594758 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-config-data\") pod \"nova-scheduler-0\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.594814 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.606540 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.607830 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-config-data\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.614383 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-scripts\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.631722 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk9d2\" (UniqueName: \"kubernetes.io/projected/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-kube-api-access-gk9d2\") pod \"nova-cell0-cell-mapping-8p47j\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.667436 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.669286 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.678604 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.690214 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.712804 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9w98\" (UniqueName: \"kubernetes.io/projected/98e35d19-7438-4960-a022-de8ab7795592-kube-api-access-g9w98\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.712858 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sct4t\" (UniqueName: \"kubernetes.io/projected/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-kube-api-access-sct4t\") pod \"nova-scheduler-0\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.712947 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-config-data\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.712999 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-config-data\") pod \"nova-scheduler-0\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.713028 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98e35d19-7438-4960-a022-de8ab7795592-logs\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.713048 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.713086 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.726058 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-config-data\") pod \"nova-scheduler-0\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.732457 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.749784 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sct4t\" (UniqueName: \"kubernetes.io/projected/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-kube-api-access-sct4t\") pod \"nova-scheduler-0\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.764585 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.767972 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b97cb559f-944lv"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.770072 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.784801 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.791357 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.794866 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.819160 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmdr6\" (UniqueName: \"kubernetes.io/projected/7ac64487-4fa7-42aa-a026-feedaaff7a94-kube-api-access-vmdr6\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.819196 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98e35d19-7438-4960-a022-de8ab7795592-logs\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.819231 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-nb\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.819251 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.827990 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b97cb559f-944lv"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.830878 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9w98\" (UniqueName: \"kubernetes.io/projected/98e35d19-7438-4960-a022-de8ab7795592-kube-api-access-g9w98\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.831017 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-sb\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.831096 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-dns-svc\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.831160 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-config-data\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.831188 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-config\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.841088 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98e35d19-7438-4960-a022-de8ab7795592-logs\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.857436 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-config-data\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.865129 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.871868 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9w98\" (UniqueName: \"kubernetes.io/projected/98e35d19-7438-4960-a022-de8ab7795592-kube-api-access-g9w98\") pod \"nova-metadata-0\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.881283 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.886669 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.887530 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.935786 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-config-data\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.935897 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-sb\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.935921 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz4jm\" (UniqueName: \"kubernetes.io/projected/6e55d4de-cf12-4f08-9995-b390b62842b6-kube-api-access-wz4jm\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.935960 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-dns-svc\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.936001 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-config\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.936037 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.936080 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmdr6\" (UniqueName: \"kubernetes.io/projected/7ac64487-4fa7-42aa-a026-feedaaff7a94-kube-api-access-vmdr6\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.936101 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-nb\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.936157 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e55d4de-cf12-4f08-9995-b390b62842b6-logs\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.936977 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-config\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.937592 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-sb\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.937819 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-nb\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.940827 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-dns-svc\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.965930 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmdr6\" (UniqueName: \"kubernetes.io/projected/7ac64487-4fa7-42aa-a026-feedaaff7a94-kube-api-access-vmdr6\") pod \"dnsmasq-dns-5b97cb559f-944lv\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.974155 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.975593 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:10 crc kubenswrapper[4822]: I1124 14:40:10.978580 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.028404 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.038468 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz4jm\" (UniqueName: \"kubernetes.io/projected/6e55d4de-cf12-4f08-9995-b390b62842b6-kube-api-access-wz4jm\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.038605 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.038638 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.038664 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.038743 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn7qb\" (UniqueName: \"kubernetes.io/projected/3ed619d2-3396-4407-bbb8-8f042307c43c-kube-api-access-pn7qb\") pod \"nova-cell1-novncproxy-0\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.038765 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e55d4de-cf12-4f08-9995-b390b62842b6-logs\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.038805 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-config-data\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.047394 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e55d4de-cf12-4f08-9995-b390b62842b6-logs\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.048386 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-config-data\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.055660 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.074782 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz4jm\" (UniqueName: \"kubernetes.io/projected/6e55d4de-cf12-4f08-9995-b390b62842b6-kube-api-access-wz4jm\") pod \"nova-api-0\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " pod="openstack/nova-api-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.141520 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.141578 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.141643 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn7qb\" (UniqueName: \"kubernetes.io/projected/3ed619d2-3396-4407-bbb8-8f042307c43c-kube-api-access-pn7qb\") pod \"nova-cell1-novncproxy-0\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.147831 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.155837 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.161029 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn7qb\" (UniqueName: \"kubernetes.io/projected/3ed619d2-3396-4407-bbb8-8f042307c43c-kube-api-access-pn7qb\") pod \"nova-cell1-novncproxy-0\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.233591 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.257470 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.331968 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.362430 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8p47j"] Nov 24 14:40:11 crc kubenswrapper[4822]: W1124 14:40:11.378838 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdf7a5a9_222a_4012_be03_5135f9bc0b1c.slice/crio-fd4b5c6c243bf13774727fefd9fe47a1eefb3fba08f225a4eba0215b4b411e06 WatchSource:0}: Error finding container fd4b5c6c243bf13774727fefd9fe47a1eefb3fba08f225a4eba0215b4b411e06: Status 404 returned error can't find the container with id fd4b5c6c243bf13774727fefd9fe47a1eefb3fba08f225a4eba0215b4b411e06 Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.417361 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8p47j" event={"ID":"bdf7a5a9-222a-4012-be03-5135f9bc0b1c","Type":"ContainerStarted","Data":"fd4b5c6c243bf13774727fefd9fe47a1eefb3fba08f225a4eba0215b4b411e06"} Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.563050 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:11 crc kubenswrapper[4822]: W1124 14:40:11.572907 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98e35d19_7438_4960_a022_de8ab7795592.slice/crio-e330913a2470b125193e6483fd2040f1fd664759b868e8fcec4270837f52a52a WatchSource:0}: Error finding container e330913a2470b125193e6483fd2040f1fd664759b868e8fcec4270837f52a52a: Status 404 returned error can't find the container with id e330913a2470b125193e6483fd2040f1fd664759b868e8fcec4270837f52a52a Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.669028 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.773769 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z2dll"] Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.777465 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.791086 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.792258 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.805111 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z2dll"] Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.840854 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b97cb559f-944lv"] Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.849100 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.873679 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-config-data\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: W1124 14:40:11.873874 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ac64487_4fa7_42aa_a026_feedaaff7a94.slice/crio-56ed15770927c4aa78df8320489697f13901a1c930d5853605f81cdbcae48136 WatchSource:0}: Error finding container 56ed15770927c4aa78df8320489697f13901a1c930d5853605f81cdbcae48136: Status 404 returned error can't find the container with id 56ed15770927c4aa78df8320489697f13901a1c930d5853605f81cdbcae48136 Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.873934 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-scripts\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.874161 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrjhw\" (UniqueName: \"kubernetes.io/projected/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-kube-api-access-lrjhw\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.874325 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.975821 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.976191 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-config-data\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.976304 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-scripts\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.976379 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrjhw\" (UniqueName: \"kubernetes.io/projected/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-kube-api-access-lrjhw\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.979415 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.981014 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-scripts\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.981889 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-config-data\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:11 crc kubenswrapper[4822]: I1124 14:40:11.996984 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrjhw\" (UniqueName: \"kubernetes.io/projected/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-kube-api-access-lrjhw\") pod \"nova-cell1-conductor-db-sync-z2dll\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.130699 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.154050 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.469526 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c","Type":"ContainerStarted","Data":"e22c7432bbe5757974db6f860f4ec947fe815ff8fbfeb206d657a82f22ec96c4"} Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.478310 4822 generic.go:334] "Generic (PLEG): container finished" podID="7ac64487-4fa7-42aa-a026-feedaaff7a94" containerID="3cff95b80e493ae40a3886e8bef4b414e4e8b5154aad25f11b406bd789437130" exitCode=0 Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.478406 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" event={"ID":"7ac64487-4fa7-42aa-a026-feedaaff7a94","Type":"ContainerDied","Data":"3cff95b80e493ae40a3886e8bef4b414e4e8b5154aad25f11b406bd789437130"} Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.478438 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" event={"ID":"7ac64487-4fa7-42aa-a026-feedaaff7a94","Type":"ContainerStarted","Data":"56ed15770927c4aa78df8320489697f13901a1c930d5853605f81cdbcae48136"} Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.490911 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8p47j" event={"ID":"bdf7a5a9-222a-4012-be03-5135f9bc0b1c","Type":"ContainerStarted","Data":"7b2a840d52fadd5f75579043238b07d839f79f0535a77d8b8c038d0af0b9770e"} Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.499341 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3ed619d2-3396-4407-bbb8-8f042307c43c","Type":"ContainerStarted","Data":"4c117e4d0974524e58449a9bb92367674bfc5c298e764d79b42f5a68563b3bdd"} Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.502575 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e55d4de-cf12-4f08-9995-b390b62842b6","Type":"ContainerStarted","Data":"7f0905d9e99b12b22b9589ba9eda22ff6e6b70e203ff6b8d72d4b07ac6e77994"} Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.504387 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"98e35d19-7438-4960-a022-de8ab7795592","Type":"ContainerStarted","Data":"e330913a2470b125193e6483fd2040f1fd664759b868e8fcec4270837f52a52a"} Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.518888 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-8p47j" podStartSLOduration=2.518870281 podStartE2EDuration="2.518870281s" podCreationTimestamp="2025-11-24 14:40:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:12.515531855 +0000 UTC m=+1249.632172332" watchObservedRunningTime="2025-11-24 14:40:12.518870281 +0000 UTC m=+1249.635510748" Nov 24 14:40:12 crc kubenswrapper[4822]: I1124 14:40:12.767553 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z2dll"] Nov 24 14:40:12 crc kubenswrapper[4822]: W1124 14:40:12.771508 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b7c7bb8_bcd0_4da3_aac0_204a07f26617.slice/crio-37846608076f601de40d5f7f28b362cef456070d5102679ff0745369990e5731 WatchSource:0}: Error finding container 37846608076f601de40d5f7f28b362cef456070d5102679ff0745369990e5731: Status 404 returned error can't find the container with id 37846608076f601de40d5f7f28b362cef456070d5102679ff0745369990e5731 Nov 24 14:40:13 crc kubenswrapper[4822]: I1124 14:40:13.525422 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z2dll" event={"ID":"8b7c7bb8-bcd0-4da3-aac0-204a07f26617","Type":"ContainerStarted","Data":"84cdfdc69581ec29622a5fc153a5198c8cecf84d476e10580993bf03c047c7f3"} Nov 24 14:40:13 crc kubenswrapper[4822]: I1124 14:40:13.525739 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z2dll" event={"ID":"8b7c7bb8-bcd0-4da3-aac0-204a07f26617","Type":"ContainerStarted","Data":"37846608076f601de40d5f7f28b362cef456070d5102679ff0745369990e5731"} Nov 24 14:40:13 crc kubenswrapper[4822]: I1124 14:40:13.528807 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" event={"ID":"7ac64487-4fa7-42aa-a026-feedaaff7a94","Type":"ContainerStarted","Data":"bb7d3198e609a9fd1f2ea6654d223909ccd890bbabc87a7ba4e1e7cf06c07825"} Nov 24 14:40:13 crc kubenswrapper[4822]: I1124 14:40:13.529149 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:13 crc kubenswrapper[4822]: I1124 14:40:13.555412 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-z2dll" podStartSLOduration=2.55538411 podStartE2EDuration="2.55538411s" podCreationTimestamp="2025-11-24 14:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:13.538738192 +0000 UTC m=+1250.655378669" watchObservedRunningTime="2025-11-24 14:40:13.55538411 +0000 UTC m=+1250.672024587" Nov 24 14:40:13 crc kubenswrapper[4822]: I1124 14:40:13.565169 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" podStartSLOduration=3.56514082 podStartE2EDuration="3.56514082s" podCreationTimestamp="2025-11-24 14:40:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:13.564549231 +0000 UTC m=+1250.681189728" watchObservedRunningTime="2025-11-24 14:40:13.56514082 +0000 UTC m=+1250.681781297" Nov 24 14:40:14 crc kubenswrapper[4822]: I1124 14:40:14.344076 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:14 crc kubenswrapper[4822]: I1124 14:40:14.360608 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.606589 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c","Type":"ContainerStarted","Data":"b8262e793b50d163df5bfc895de7dee4a824cddd12672e0ca3469bb4cb72478f"} Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.608653 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3ed619d2-3396-4407-bbb8-8f042307c43c","Type":"ContainerStarted","Data":"e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c"} Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.608851 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3ed619d2-3396-4407-bbb8-8f042307c43c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c" gracePeriod=30 Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.618282 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e55d4de-cf12-4f08-9995-b390b62842b6","Type":"ContainerStarted","Data":"68203306cc053545db28dc12b820d204bf4b6b3d17b43211903ae7540f24bfdb"} Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.618322 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e55d4de-cf12-4f08-9995-b390b62842b6","Type":"ContainerStarted","Data":"d9ce46ddce6957a4cbccfd3b29b2280a7a8c5722240266bb48cc308f76559e11"} Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.625660 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"98e35d19-7438-4960-a022-de8ab7795592","Type":"ContainerStarted","Data":"26779d09b625930146039c7321ef57a5b53f2379d8c7598e543b780d44ee6b21"} Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.625726 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"98e35d19-7438-4960-a022-de8ab7795592","Type":"ContainerStarted","Data":"401b5877c1423f8560c033943a64d0adb1c5b6653cc4fffe0855c54fbf17445b"} Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.625903 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="98e35d19-7438-4960-a022-de8ab7795592" containerName="nova-metadata-log" containerID="cri-o://401b5877c1423f8560c033943a64d0adb1c5b6653cc4fffe0855c54fbf17445b" gracePeriod=30 Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.626230 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="98e35d19-7438-4960-a022-de8ab7795592" containerName="nova-metadata-metadata" containerID="cri-o://26779d09b625930146039c7321ef57a5b53f2379d8c7598e543b780d44ee6b21" gracePeriod=30 Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.639636 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.598822765 podStartE2EDuration="6.639615849s" podCreationTimestamp="2025-11-24 14:40:10 +0000 UTC" firstStartedPulling="2025-11-24 14:40:11.67765362 +0000 UTC m=+1248.794294087" lastFinishedPulling="2025-11-24 14:40:15.718446694 +0000 UTC m=+1252.835087171" observedRunningTime="2025-11-24 14:40:16.627526945 +0000 UTC m=+1253.744167422" watchObservedRunningTime="2025-11-24 14:40:16.639615849 +0000 UTC m=+1253.756256316" Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.660833 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8161469820000002 podStartE2EDuration="6.66081274s" podCreationTimestamp="2025-11-24 14:40:10 +0000 UTC" firstStartedPulling="2025-11-24 14:40:11.87390146 +0000 UTC m=+1248.990541937" lastFinishedPulling="2025-11-24 14:40:15.718567218 +0000 UTC m=+1252.835207695" observedRunningTime="2025-11-24 14:40:16.65704252 +0000 UTC m=+1253.773683027" watchObservedRunningTime="2025-11-24 14:40:16.66081274 +0000 UTC m=+1253.777453217" Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.700101 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.560081326 podStartE2EDuration="6.700072564s" podCreationTimestamp="2025-11-24 14:40:10 +0000 UTC" firstStartedPulling="2025-11-24 14:40:11.57794168 +0000 UTC m=+1248.694582147" lastFinishedPulling="2025-11-24 14:40:15.717932908 +0000 UTC m=+1252.834573385" observedRunningTime="2025-11-24 14:40:16.687457035 +0000 UTC m=+1253.804097532" watchObservedRunningTime="2025-11-24 14:40:16.700072564 +0000 UTC m=+1253.816713041" Nov 24 14:40:16 crc kubenswrapper[4822]: I1124 14:40:16.713109 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.217106639 podStartE2EDuration="6.713087136s" podCreationTimestamp="2025-11-24 14:40:10 +0000 UTC" firstStartedPulling="2025-11-24 14:40:12.222387035 +0000 UTC m=+1249.339027512" lastFinishedPulling="2025-11-24 14:40:15.718367532 +0000 UTC m=+1252.835008009" observedRunningTime="2025-11-24 14:40:16.707556731 +0000 UTC m=+1253.824197238" watchObservedRunningTime="2025-11-24 14:40:16.713087136 +0000 UTC m=+1253.829727623" Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.646067 4822 generic.go:334] "Generic (PLEG): container finished" podID="98e35d19-7438-4960-a022-de8ab7795592" containerID="26779d09b625930146039c7321ef57a5b53f2379d8c7598e543b780d44ee6b21" exitCode=0 Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.646333 4822 generic.go:334] "Generic (PLEG): container finished" podID="98e35d19-7438-4960-a022-de8ab7795592" containerID="401b5877c1423f8560c033943a64d0adb1c5b6653cc4fffe0855c54fbf17445b" exitCode=143 Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.646106 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"98e35d19-7438-4960-a022-de8ab7795592","Type":"ContainerDied","Data":"26779d09b625930146039c7321ef57a5b53f2379d8c7598e543b780d44ee6b21"} Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.646389 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"98e35d19-7438-4960-a022-de8ab7795592","Type":"ContainerDied","Data":"401b5877c1423f8560c033943a64d0adb1c5b6653cc4fffe0855c54fbf17445b"} Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.770008 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.830087 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-config-data\") pod \"98e35d19-7438-4960-a022-de8ab7795592\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.830490 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-combined-ca-bundle\") pod \"98e35d19-7438-4960-a022-de8ab7795592\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.830581 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98e35d19-7438-4960-a022-de8ab7795592-logs\") pod \"98e35d19-7438-4960-a022-de8ab7795592\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.830721 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9w98\" (UniqueName: \"kubernetes.io/projected/98e35d19-7438-4960-a022-de8ab7795592-kube-api-access-g9w98\") pod \"98e35d19-7438-4960-a022-de8ab7795592\" (UID: \"98e35d19-7438-4960-a022-de8ab7795592\") " Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.831050 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98e35d19-7438-4960-a022-de8ab7795592-logs" (OuterVolumeSpecName: "logs") pod "98e35d19-7438-4960-a022-de8ab7795592" (UID: "98e35d19-7438-4960-a022-de8ab7795592"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.831714 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98e35d19-7438-4960-a022-de8ab7795592-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.856404 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98e35d19-7438-4960-a022-de8ab7795592-kube-api-access-g9w98" (OuterVolumeSpecName: "kube-api-access-g9w98") pod "98e35d19-7438-4960-a022-de8ab7795592" (UID: "98e35d19-7438-4960-a022-de8ab7795592"). InnerVolumeSpecName "kube-api-access-g9w98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.873647 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98e35d19-7438-4960-a022-de8ab7795592" (UID: "98e35d19-7438-4960-a022-de8ab7795592"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.881334 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-config-data" (OuterVolumeSpecName: "config-data") pod "98e35d19-7438-4960-a022-de8ab7795592" (UID: "98e35d19-7438-4960-a022-de8ab7795592"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.933598 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.933634 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9w98\" (UniqueName: \"kubernetes.io/projected/98e35d19-7438-4960-a022-de8ab7795592-kube-api-access-g9w98\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:17 crc kubenswrapper[4822]: I1124 14:40:17.933645 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e35d19-7438-4960-a022-de8ab7795592-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.662272 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"98e35d19-7438-4960-a022-de8ab7795592","Type":"ContainerDied","Data":"e330913a2470b125193e6483fd2040f1fd664759b868e8fcec4270837f52a52a"} Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.662327 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.662339 4822 scope.go:117] "RemoveContainer" containerID="26779d09b625930146039c7321ef57a5b53f2379d8c7598e543b780d44ee6b21" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.700984 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.713108 4822 scope.go:117] "RemoveContainer" containerID="401b5877c1423f8560c033943a64d0adb1c5b6653cc4fffe0855c54fbf17445b" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.713346 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.723348 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:18 crc kubenswrapper[4822]: E1124 14:40:18.723874 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98e35d19-7438-4960-a022-de8ab7795592" containerName="nova-metadata-log" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.723893 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="98e35d19-7438-4960-a022-de8ab7795592" containerName="nova-metadata-log" Nov 24 14:40:18 crc kubenswrapper[4822]: E1124 14:40:18.723939 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98e35d19-7438-4960-a022-de8ab7795592" containerName="nova-metadata-metadata" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.723946 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="98e35d19-7438-4960-a022-de8ab7795592" containerName="nova-metadata-metadata" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.724161 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="98e35d19-7438-4960-a022-de8ab7795592" containerName="nova-metadata-metadata" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.724189 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="98e35d19-7438-4960-a022-de8ab7795592" containerName="nova-metadata-log" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.725561 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.728694 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.728857 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.745136 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.852921 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.853262 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.853325 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d51d6c-85ba-42da-9d90-e25359ee1db4-logs\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.853925 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9czts\" (UniqueName: \"kubernetes.io/projected/63d51d6c-85ba-42da-9d90-e25359ee1db4-kube-api-access-9czts\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.854027 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-config-data\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.956602 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.956691 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d51d6c-85ba-42da-9d90-e25359ee1db4-logs\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.956882 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9czts\" (UniqueName: \"kubernetes.io/projected/63d51d6c-85ba-42da-9d90-e25359ee1db4-kube-api-access-9czts\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.956915 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-config-data\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.956960 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.957378 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d51d6c-85ba-42da-9d90-e25359ee1db4-logs\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.960577 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.963015 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-config-data\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.964460 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:18 crc kubenswrapper[4822]: I1124 14:40:18.976254 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9czts\" (UniqueName: \"kubernetes.io/projected/63d51d6c-85ba-42da-9d90-e25359ee1db4-kube-api-access-9czts\") pod \"nova-metadata-0\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " pod="openstack/nova-metadata-0" Nov 24 14:40:19 crc kubenswrapper[4822]: I1124 14:40:19.050784 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:19 crc kubenswrapper[4822]: I1124 14:40:19.626367 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:19 crc kubenswrapper[4822]: I1124 14:40:19.674369 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63d51d6c-85ba-42da-9d90-e25359ee1db4","Type":"ContainerStarted","Data":"e7d618bb09fe8d7a4eb3903796b4caaf77724b8f2614a6ad9aab08ad92c6815c"} Nov 24 14:40:19 crc kubenswrapper[4822]: I1124 14:40:19.720578 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98e35d19-7438-4960-a022-de8ab7795592" path="/var/lib/kubelet/pods/98e35d19-7438-4960-a022-de8ab7795592/volumes" Nov 24 14:40:20 crc kubenswrapper[4822]: I1124 14:40:20.687569 4822 generic.go:334] "Generic (PLEG): container finished" podID="bdf7a5a9-222a-4012-be03-5135f9bc0b1c" containerID="7b2a840d52fadd5f75579043238b07d839f79f0535a77d8b8c038d0af0b9770e" exitCode=0 Nov 24 14:40:20 crc kubenswrapper[4822]: I1124 14:40:20.687632 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8p47j" event={"ID":"bdf7a5a9-222a-4012-be03-5135f9bc0b1c","Type":"ContainerDied","Data":"7b2a840d52fadd5f75579043238b07d839f79f0535a77d8b8c038d0af0b9770e"} Nov 24 14:40:20 crc kubenswrapper[4822]: I1124 14:40:20.690360 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63d51d6c-85ba-42da-9d90-e25359ee1db4","Type":"ContainerStarted","Data":"a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb"} Nov 24 14:40:20 crc kubenswrapper[4822]: I1124 14:40:20.690421 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63d51d6c-85ba-42da-9d90-e25359ee1db4","Type":"ContainerStarted","Data":"00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb"} Nov 24 14:40:20 crc kubenswrapper[4822]: I1124 14:40:20.735976 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7359541529999998 podStartE2EDuration="2.735954153s" podCreationTimestamp="2025-11-24 14:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:20.729168658 +0000 UTC m=+1257.845809145" watchObservedRunningTime="2025-11-24 14:40:20.735954153 +0000 UTC m=+1257.852594620" Nov 24 14:40:20 crc kubenswrapper[4822]: I1124 14:40:20.887128 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:40:20 crc kubenswrapper[4822]: I1124 14:40:20.887493 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 14:40:20 crc kubenswrapper[4822]: I1124 14:40:20.941403 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.236515 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.259179 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.259275 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.333652 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.346467 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c56c4ff7-2xkkd"] Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.346774 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" podUID="f6d5fe88-9544-489a-8fab-1b0ed7781043" containerName="dnsmasq-dns" containerID="cri-o://33deb1cb2c718cba129e94fa4cdeebe41fe8b60a7058899d10d63d556af393f3" gracePeriod=10 Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.723023 4822 generic.go:334] "Generic (PLEG): container finished" podID="f6d5fe88-9544-489a-8fab-1b0ed7781043" containerID="33deb1cb2c718cba129e94fa4cdeebe41fe8b60a7058899d10d63d556af393f3" exitCode=0 Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.724879 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" event={"ID":"f6d5fe88-9544-489a-8fab-1b0ed7781043","Type":"ContainerDied","Data":"33deb1cb2c718cba129e94fa4cdeebe41fe8b60a7058899d10d63d556af393f3"} Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.738618 4822 generic.go:334] "Generic (PLEG): container finished" podID="8b7c7bb8-bcd0-4da3-aac0-204a07f26617" containerID="84cdfdc69581ec29622a5fc153a5198c8cecf84d476e10580993bf03c047c7f3" exitCode=0 Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.738925 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z2dll" event={"ID":"8b7c7bb8-bcd0-4da3-aac0-204a07f26617","Type":"ContainerDied","Data":"84cdfdc69581ec29622a5fc153a5198c8cecf84d476e10580993bf03c047c7f3"} Nov 24 14:40:21 crc kubenswrapper[4822]: I1124 14:40:21.776470 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.060597 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.130719 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-nb\") pod \"f6d5fe88-9544-489a-8fab-1b0ed7781043\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.130793 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-sb\") pod \"f6d5fe88-9544-489a-8fab-1b0ed7781043\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.130865 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-config\") pod \"f6d5fe88-9544-489a-8fab-1b0ed7781043\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.130890 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-dns-svc\") pod \"f6d5fe88-9544-489a-8fab-1b0ed7781043\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.131058 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bdxt\" (UniqueName: \"kubernetes.io/projected/f6d5fe88-9544-489a-8fab-1b0ed7781043-kube-api-access-6bdxt\") pod \"f6d5fe88-9544-489a-8fab-1b0ed7781043\" (UID: \"f6d5fe88-9544-489a-8fab-1b0ed7781043\") " Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.149499 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6d5fe88-9544-489a-8fab-1b0ed7781043-kube-api-access-6bdxt" (OuterVolumeSpecName: "kube-api-access-6bdxt") pod "f6d5fe88-9544-489a-8fab-1b0ed7781043" (UID: "f6d5fe88-9544-489a-8fab-1b0ed7781043"). InnerVolumeSpecName "kube-api-access-6bdxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.193516 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f6d5fe88-9544-489a-8fab-1b0ed7781043" (UID: "f6d5fe88-9544-489a-8fab-1b0ed7781043"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.234570 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.234878 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bdxt\" (UniqueName: \"kubernetes.io/projected/f6d5fe88-9544-489a-8fab-1b0ed7781043-kube-api-access-6bdxt\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.239972 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-config" (OuterVolumeSpecName: "config") pod "f6d5fe88-9544-489a-8fab-1b0ed7781043" (UID: "f6d5fe88-9544-489a-8fab-1b0ed7781043"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.254612 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.264619 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f6d5fe88-9544-489a-8fab-1b0ed7781043" (UID: "f6d5fe88-9544-489a-8fab-1b0ed7781043"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.306732 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f6d5fe88-9544-489a-8fab-1b0ed7781043" (UID: "f6d5fe88-9544-489a-8fab-1b0ed7781043"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.336304 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-config-data\") pod \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.336431 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-scripts\") pod \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.336569 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-combined-ca-bundle\") pod \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.336670 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk9d2\" (UniqueName: \"kubernetes.io/projected/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-kube-api-access-gk9d2\") pod \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\" (UID: \"bdf7a5a9-222a-4012-be03-5135f9bc0b1c\") " Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.337128 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.337145 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.337154 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d5fe88-9544-489a-8fab-1b0ed7781043-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.340739 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-scripts" (OuterVolumeSpecName: "scripts") pod "bdf7a5a9-222a-4012-be03-5135f9bc0b1c" (UID: "bdf7a5a9-222a-4012-be03-5135f9bc0b1c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.346090 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-kube-api-access-gk9d2" (OuterVolumeSpecName: "kube-api-access-gk9d2") pod "bdf7a5a9-222a-4012-be03-5135f9bc0b1c" (UID: "bdf7a5a9-222a-4012-be03-5135f9bc0b1c"). InnerVolumeSpecName "kube-api-access-gk9d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.346610 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.211:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.346976 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.211:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.364418 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdf7a5a9-222a-4012-be03-5135f9bc0b1c" (UID: "bdf7a5a9-222a-4012-be03-5135f9bc0b1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.365133 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-config-data" (OuterVolumeSpecName: "config-data") pod "bdf7a5a9-222a-4012-be03-5135f9bc0b1c" (UID: "bdf7a5a9-222a-4012-be03-5135f9bc0b1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.441273 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.441304 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk9d2\" (UniqueName: \"kubernetes.io/projected/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-kube-api-access-gk9d2\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.441316 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.441325 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf7a5a9-222a-4012-be03-5135f9bc0b1c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.756091 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8p47j" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.756488 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8p47j" event={"ID":"bdf7a5a9-222a-4012-be03-5135f9bc0b1c","Type":"ContainerDied","Data":"fd4b5c6c243bf13774727fefd9fe47a1eefb3fba08f225a4eba0215b4b411e06"} Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.756536 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd4b5c6c243bf13774727fefd9fe47a1eefb3fba08f225a4eba0215b4b411e06" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.759439 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" event={"ID":"f6d5fe88-9544-489a-8fab-1b0ed7781043","Type":"ContainerDied","Data":"34b3562521fac0f0d263b7b2b9937419f783e6bdd0d6154421b470ffe659f8bd"} Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.759478 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c56c4ff7-2xkkd" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.759517 4822 scope.go:117] "RemoveContainer" containerID="33deb1cb2c718cba129e94fa4cdeebe41fe8b60a7058899d10d63d556af393f3" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.807298 4822 scope.go:117] "RemoveContainer" containerID="06725a3b562a7d0e390d3800840fb5829929c539bd0e1e05e34ca42a8766b748" Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.823294 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c56c4ff7-2xkkd"] Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.836503 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c56c4ff7-2xkkd"] Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.921139 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.921423 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-log" containerID="cri-o://d9ce46ddce6957a4cbccfd3b29b2280a7a8c5722240266bb48cc308f76559e11" gracePeriod=30 Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.921879 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-api" containerID="cri-o://68203306cc053545db28dc12b820d204bf4b6b3d17b43211903ae7540f24bfdb" gracePeriod=30 Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.944090 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.977791 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.978142 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerName="nova-metadata-log" containerID="cri-o://00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb" gracePeriod=30 Nov 24 14:40:22 crc kubenswrapper[4822]: I1124 14:40:22.978661 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerName="nova-metadata-metadata" containerID="cri-o://a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb" gracePeriod=30 Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.269379 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.365649 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-combined-ca-bundle\") pod \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.365755 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrjhw\" (UniqueName: \"kubernetes.io/projected/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-kube-api-access-lrjhw\") pod \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.366170 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-config-data\") pod \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.366275 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-scripts\") pod \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\" (UID: \"8b7c7bb8-bcd0-4da3-aac0-204a07f26617\") " Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.373328 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-kube-api-access-lrjhw" (OuterVolumeSpecName: "kube-api-access-lrjhw") pod "8b7c7bb8-bcd0-4da3-aac0-204a07f26617" (UID: "8b7c7bb8-bcd0-4da3-aac0-204a07f26617"). InnerVolumeSpecName "kube-api-access-lrjhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.380563 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-scripts" (OuterVolumeSpecName: "scripts") pod "8b7c7bb8-bcd0-4da3-aac0-204a07f26617" (UID: "8b7c7bb8-bcd0-4da3-aac0-204a07f26617"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.396496 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b7c7bb8-bcd0-4da3-aac0-204a07f26617" (UID: "8b7c7bb8-bcd0-4da3-aac0-204a07f26617"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.408517 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-config-data" (OuterVolumeSpecName: "config-data") pod "8b7c7bb8-bcd0-4da3-aac0-204a07f26617" (UID: "8b7c7bb8-bcd0-4da3-aac0-204a07f26617"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.469856 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.469924 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.469940 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.469953 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrjhw\" (UniqueName: \"kubernetes.io/projected/8b7c7bb8-bcd0-4da3-aac0-204a07f26617-kube-api-access-lrjhw\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.717064 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.717518 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6d5fe88-9544-489a-8fab-1b0ed7781043" path="/var/lib/kubelet/pods/f6d5fe88-9544-489a-8fab-1b0ed7781043/volumes" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.775365 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-config-data\") pod \"63d51d6c-85ba-42da-9d90-e25359ee1db4\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.775491 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d51d6c-85ba-42da-9d90-e25359ee1db4-logs\") pod \"63d51d6c-85ba-42da-9d90-e25359ee1db4\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.775574 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9czts\" (UniqueName: \"kubernetes.io/projected/63d51d6c-85ba-42da-9d90-e25359ee1db4-kube-api-access-9czts\") pod \"63d51d6c-85ba-42da-9d90-e25359ee1db4\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.775662 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-combined-ca-bundle\") pod \"63d51d6c-85ba-42da-9d90-e25359ee1db4\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.775720 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-nova-metadata-tls-certs\") pod \"63d51d6c-85ba-42da-9d90-e25359ee1db4\" (UID: \"63d51d6c-85ba-42da-9d90-e25359ee1db4\") " Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.777051 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63d51d6c-85ba-42da-9d90-e25359ee1db4-logs" (OuterVolumeSpecName: "logs") pod "63d51d6c-85ba-42da-9d90-e25359ee1db4" (UID: "63d51d6c-85ba-42da-9d90-e25359ee1db4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.777726 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d51d6c-85ba-42da-9d90-e25359ee1db4-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.797991 4822 generic.go:334] "Generic (PLEG): container finished" podID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerID="d9ce46ddce6957a4cbccfd3b29b2280a7a8c5722240266bb48cc308f76559e11" exitCode=143 Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.798064 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e55d4de-cf12-4f08-9995-b390b62842b6","Type":"ContainerDied","Data":"d9ce46ddce6957a4cbccfd3b29b2280a7a8c5722240266bb48cc308f76559e11"} Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.798450 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d51d6c-85ba-42da-9d90-e25359ee1db4-kube-api-access-9czts" (OuterVolumeSpecName: "kube-api-access-9czts") pod "63d51d6c-85ba-42da-9d90-e25359ee1db4" (UID: "63d51d6c-85ba-42da-9d90-e25359ee1db4"). InnerVolumeSpecName "kube-api-access-9czts". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.804150 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z2dll" event={"ID":"8b7c7bb8-bcd0-4da3-aac0-204a07f26617","Type":"ContainerDied","Data":"37846608076f601de40d5f7f28b362cef456070d5102679ff0745369990e5731"} Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.804291 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37846608076f601de40d5f7f28b362cef456070d5102679ff0745369990e5731" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.804410 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z2dll" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.810836 4822 generic.go:334] "Generic (PLEG): container finished" podID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerID="a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb" exitCode=0 Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.810950 4822 generic.go:334] "Generic (PLEG): container finished" podID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerID="00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb" exitCode=143 Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.811184 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" containerName="nova-scheduler-scheduler" containerID="cri-o://b8262e793b50d163df5bfc895de7dee4a824cddd12672e0ca3469bb4cb72478f" gracePeriod=30 Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.811798 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.813674 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63d51d6c-85ba-42da-9d90-e25359ee1db4","Type":"ContainerDied","Data":"a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb"} Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.813791 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63d51d6c-85ba-42da-9d90-e25359ee1db4","Type":"ContainerDied","Data":"00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb"} Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.813855 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63d51d6c-85ba-42da-9d90-e25359ee1db4","Type":"ContainerDied","Data":"e7d618bb09fe8d7a4eb3903796b4caaf77724b8f2614a6ad9aab08ad92c6815c"} Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.814006 4822 scope.go:117] "RemoveContainer" containerID="a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.851432 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-config-data" (OuterVolumeSpecName: "config-data") pod "63d51d6c-85ba-42da-9d90-e25359ee1db4" (UID: "63d51d6c-85ba-42da-9d90-e25359ee1db4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.854304 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:40:23 crc kubenswrapper[4822]: E1124 14:40:23.854706 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d5fe88-9544-489a-8fab-1b0ed7781043" containerName="init" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.854720 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d5fe88-9544-489a-8fab-1b0ed7781043" containerName="init" Nov 24 14:40:23 crc kubenswrapper[4822]: E1124 14:40:23.854728 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf7a5a9-222a-4012-be03-5135f9bc0b1c" containerName="nova-manage" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.854734 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf7a5a9-222a-4012-be03-5135f9bc0b1c" containerName="nova-manage" Nov 24 14:40:23 crc kubenswrapper[4822]: E1124 14:40:23.854753 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b7c7bb8-bcd0-4da3-aac0-204a07f26617" containerName="nova-cell1-conductor-db-sync" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.854760 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b7c7bb8-bcd0-4da3-aac0-204a07f26617" containerName="nova-cell1-conductor-db-sync" Nov 24 14:40:23 crc kubenswrapper[4822]: E1124 14:40:23.854769 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerName="nova-metadata-log" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.854775 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerName="nova-metadata-log" Nov 24 14:40:23 crc kubenswrapper[4822]: E1124 14:40:23.854790 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerName="nova-metadata-metadata" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.854795 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerName="nova-metadata-metadata" Nov 24 14:40:23 crc kubenswrapper[4822]: E1124 14:40:23.854814 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d5fe88-9544-489a-8fab-1b0ed7781043" containerName="dnsmasq-dns" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.854820 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d5fe88-9544-489a-8fab-1b0ed7781043" containerName="dnsmasq-dns" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.855023 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerName="nova-metadata-metadata" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.855037 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d5fe88-9544-489a-8fab-1b0ed7781043" containerName="dnsmasq-dns" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.855051 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d51d6c-85ba-42da-9d90-e25359ee1db4" containerName="nova-metadata-log" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.855069 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b7c7bb8-bcd0-4da3-aac0-204a07f26617" containerName="nova-cell1-conductor-db-sync" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.855080 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf7a5a9-222a-4012-be03-5135f9bc0b1c" containerName="nova-manage" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.855805 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.858706 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.869599 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63d51d6c-85ba-42da-9d90-e25359ee1db4" (UID: "63d51d6c-85ba-42da-9d90-e25359ee1db4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.870522 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.878879 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2336f65-386b-401e-b723-abf4666c08e4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b2336f65-386b-401e-b723-abf4666c08e4\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.903583 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2336f65-386b-401e-b723-abf4666c08e4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b2336f65-386b-401e-b723-abf4666c08e4\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.903694 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqs88\" (UniqueName: \"kubernetes.io/projected/b2336f65-386b-401e-b723-abf4666c08e4-kube-api-access-cqs88\") pod \"nova-cell1-conductor-0\" (UID: \"b2336f65-386b-401e-b723-abf4666c08e4\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.904384 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9czts\" (UniqueName: \"kubernetes.io/projected/63d51d6c-85ba-42da-9d90-e25359ee1db4-kube-api-access-9czts\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.904415 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.904425 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.934484 4822 scope.go:117] "RemoveContainer" containerID="00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.951752 4822 scope.go:117] "RemoveContainer" containerID="a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.951749 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "63d51d6c-85ba-42da-9d90-e25359ee1db4" (UID: "63d51d6c-85ba-42da-9d90-e25359ee1db4"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:23 crc kubenswrapper[4822]: E1124 14:40:23.952425 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb\": container with ID starting with a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb not found: ID does not exist" containerID="a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.952463 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb"} err="failed to get container status \"a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb\": rpc error: code = NotFound desc = could not find container \"a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb\": container with ID starting with a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb not found: ID does not exist" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.952488 4822 scope.go:117] "RemoveContainer" containerID="00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb" Nov 24 14:40:23 crc kubenswrapper[4822]: E1124 14:40:23.952785 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb\": container with ID starting with 00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb not found: ID does not exist" containerID="00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.952801 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb"} err="failed to get container status \"00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb\": rpc error: code = NotFound desc = could not find container \"00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb\": container with ID starting with 00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb not found: ID does not exist" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.952814 4822 scope.go:117] "RemoveContainer" containerID="a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.952982 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb"} err="failed to get container status \"a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb\": rpc error: code = NotFound desc = could not find container \"a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb\": container with ID starting with a8c7d706c0d6326c4c0828d1f8645afc68356f14456bd57d1086e499f277cfeb not found: ID does not exist" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.952999 4822 scope.go:117] "RemoveContainer" containerID="00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb" Nov 24 14:40:23 crc kubenswrapper[4822]: I1124 14:40:23.953165 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb"} err="failed to get container status \"00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb\": rpc error: code = NotFound desc = could not find container \"00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb\": container with ID starting with 00b2e10d42b5d25072e1b68a15c1a955bbd37fea671542a7b45ab9f6cd7cfafb not found: ID does not exist" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.006830 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2336f65-386b-401e-b723-abf4666c08e4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b2336f65-386b-401e-b723-abf4666c08e4\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.006929 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2336f65-386b-401e-b723-abf4666c08e4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b2336f65-386b-401e-b723-abf4666c08e4\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.006958 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqs88\" (UniqueName: \"kubernetes.io/projected/b2336f65-386b-401e-b723-abf4666c08e4-kube-api-access-cqs88\") pod \"nova-cell1-conductor-0\" (UID: \"b2336f65-386b-401e-b723-abf4666c08e4\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.007747 4822 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d51d6c-85ba-42da-9d90-e25359ee1db4-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.012024 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2336f65-386b-401e-b723-abf4666c08e4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b2336f65-386b-401e-b723-abf4666c08e4\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.012858 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2336f65-386b-401e-b723-abf4666c08e4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b2336f65-386b-401e-b723-abf4666c08e4\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.027678 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqs88\" (UniqueName: \"kubernetes.io/projected/b2336f65-386b-401e-b723-abf4666c08e4-kube-api-access-cqs88\") pod \"nova-cell1-conductor-0\" (UID: \"b2336f65-386b-401e-b723-abf4666c08e4\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.141747 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.153783 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.169661 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.171619 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.173485 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.174322 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.185666 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.211567 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.211610 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2h52\" (UniqueName: \"kubernetes.io/projected/0962a558-6bb1-44fe-953a-1e4f717b175e-kube-api-access-f2h52\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.211676 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.211706 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0962a558-6bb1-44fe-953a-1e4f717b175e-logs\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.211726 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-config-data\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.227643 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.312926 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.313004 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2h52\" (UniqueName: \"kubernetes.io/projected/0962a558-6bb1-44fe-953a-1e4f717b175e-kube-api-access-f2h52\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.313079 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.313111 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0962a558-6bb1-44fe-953a-1e4f717b175e-logs\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.313130 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-config-data\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.313750 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0962a558-6bb1-44fe-953a-1e4f717b175e-logs\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.322085 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.326913 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-config-data\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.327376 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.350855 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2h52\" (UniqueName: \"kubernetes.io/projected/0962a558-6bb1-44fe-953a-1e4f717b175e-kube-api-access-f2h52\") pod \"nova-metadata-0\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.497869 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.769769 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:40:24 crc kubenswrapper[4822]: W1124 14:40:24.776431 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2336f65_386b_401e_b723_abf4666c08e4.slice/crio-eb9a8865e11cbb295d659b804237b7e556ecf6d515ecce653c1b13293169eb8b WatchSource:0}: Error finding container eb9a8865e11cbb295d659b804237b7e556ecf6d515ecce653c1b13293169eb8b: Status 404 returned error can't find the container with id eb9a8865e11cbb295d659b804237b7e556ecf6d515ecce653c1b13293169eb8b Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.831689 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b2336f65-386b-401e-b723-abf4666c08e4","Type":"ContainerStarted","Data":"eb9a8865e11cbb295d659b804237b7e556ecf6d515ecce653c1b13293169eb8b"} Nov 24 14:40:24 crc kubenswrapper[4822]: I1124 14:40:24.998068 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:40:25 crc kubenswrapper[4822]: W1124 14:40:25.011356 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0962a558_6bb1_44fe_953a_1e4f717b175e.slice/crio-8d7df195c3a162c730126881ed828cb4b803dc665532e26e7ff25a241ae2d5f3 WatchSource:0}: Error finding container 8d7df195c3a162c730126881ed828cb4b803dc665532e26e7ff25a241ae2d5f3: Status 404 returned error can't find the container with id 8d7df195c3a162c730126881ed828cb4b803dc665532e26e7ff25a241ae2d5f3 Nov 24 14:40:25 crc kubenswrapper[4822]: I1124 14:40:25.719033 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63d51d6c-85ba-42da-9d90-e25359ee1db4" path="/var/lib/kubelet/pods/63d51d6c-85ba-42da-9d90-e25359ee1db4/volumes" Nov 24 14:40:25 crc kubenswrapper[4822]: I1124 14:40:25.843746 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0962a558-6bb1-44fe-953a-1e4f717b175e","Type":"ContainerStarted","Data":"e6533dec0b4a77a1cc2e167c2758c732b27d3561356b0dd94e42324fb97a8c32"} Nov 24 14:40:25 crc kubenswrapper[4822]: I1124 14:40:25.843802 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0962a558-6bb1-44fe-953a-1e4f717b175e","Type":"ContainerStarted","Data":"63b708115911fa584ad96a7bf285d6763f12a71f178e4989618107f79a33305d"} Nov 24 14:40:25 crc kubenswrapper[4822]: I1124 14:40:25.843817 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0962a558-6bb1-44fe-953a-1e4f717b175e","Type":"ContainerStarted","Data":"8d7df195c3a162c730126881ed828cb4b803dc665532e26e7ff25a241ae2d5f3"} Nov 24 14:40:25 crc kubenswrapper[4822]: I1124 14:40:25.846826 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b2336f65-386b-401e-b723-abf4666c08e4","Type":"ContainerStarted","Data":"c69ee8b939c77589673d0291c1e869abd015b321d846ab0e3385f6f0f95af1bf"} Nov 24 14:40:25 crc kubenswrapper[4822]: I1124 14:40:25.847067 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:25 crc kubenswrapper[4822]: I1124 14:40:25.876366 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.876344676 podStartE2EDuration="1.876344676s" podCreationTimestamp="2025-11-24 14:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:25.870623935 +0000 UTC m=+1262.987264432" watchObservedRunningTime="2025-11-24 14:40:25.876344676 +0000 UTC m=+1262.992985143" Nov 24 14:40:25 crc kubenswrapper[4822]: E1124 14:40:25.893686 4822 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b8262e793b50d163df5bfc895de7dee4a824cddd12672e0ca3469bb4cb72478f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:40:25 crc kubenswrapper[4822]: E1124 14:40:25.895176 4822 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b8262e793b50d163df5bfc895de7dee4a824cddd12672e0ca3469bb4cb72478f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:40:25 crc kubenswrapper[4822]: E1124 14:40:25.897530 4822 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b8262e793b50d163df5bfc895de7dee4a824cddd12672e0ca3469bb4cb72478f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:40:25 crc kubenswrapper[4822]: E1124 14:40:25.897619 4822 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" containerName="nova-scheduler-scheduler" Nov 24 14:40:25 crc kubenswrapper[4822]: I1124 14:40:25.902650 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.90262289 podStartE2EDuration="2.90262289s" podCreationTimestamp="2025-11-24 14:40:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:25.891685153 +0000 UTC m=+1263.008325630" watchObservedRunningTime="2025-11-24 14:40:25.90262289 +0000 UTC m=+1263.019263377" Nov 24 14:40:27 crc kubenswrapper[4822]: I1124 14:40:27.875890 4822 generic.go:334] "Generic (PLEG): container finished" podID="ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" containerID="b8262e793b50d163df5bfc895de7dee4a824cddd12672e0ca3469bb4cb72478f" exitCode=0 Nov 24 14:40:27 crc kubenswrapper[4822]: I1124 14:40:27.876080 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c","Type":"ContainerDied","Data":"b8262e793b50d163df5bfc895de7dee4a824cddd12672e0ca3469bb4cb72478f"} Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.213761 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.324894 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sct4t\" (UniqueName: \"kubernetes.io/projected/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-kube-api-access-sct4t\") pod \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.325040 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-config-data\") pod \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.325140 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-combined-ca-bundle\") pod \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\" (UID: \"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c\") " Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.357557 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-kube-api-access-sct4t" (OuterVolumeSpecName: "kube-api-access-sct4t") pod "ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" (UID: "ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c"). InnerVolumeSpecName "kube-api-access-sct4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.362678 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-config-data" (OuterVolumeSpecName: "config-data") pod "ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" (UID: "ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.405374 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" (UID: "ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.426932 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.426964 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.426975 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sct4t\" (UniqueName: \"kubernetes.io/projected/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c-kube-api-access-sct4t\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.906909 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c","Type":"ContainerDied","Data":"e22c7432bbe5757974db6f860f4ec947fe815ff8fbfeb206d657a82f22ec96c4"} Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.907267 4822 scope.go:117] "RemoveContainer" containerID="b8262e793b50d163df5bfc895de7dee4a824cddd12672e0ca3469bb4cb72478f" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.907447 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.913705 4822 generic.go:334] "Generic (PLEG): container finished" podID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerID="68203306cc053545db28dc12b820d204bf4b6b3d17b43211903ae7540f24bfdb" exitCode=0 Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.913746 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e55d4de-cf12-4f08-9995-b390b62842b6","Type":"ContainerDied","Data":"68203306cc053545db28dc12b820d204bf4b6b3d17b43211903ae7540f24bfdb"} Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.913774 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e55d4de-cf12-4f08-9995-b390b62842b6","Type":"ContainerDied","Data":"7f0905d9e99b12b22b9589ba9eda22ff6e6b70e203ff6b8d72d4b07ac6e77994"} Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.913786 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f0905d9e99b12b22b9589ba9eda22ff6e6b70e203ff6b8d72d4b07ac6e77994" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.974101 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:28 crc kubenswrapper[4822]: I1124 14:40:28.996630 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.008441 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.026764 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:40:29 crc kubenswrapper[4822]: E1124 14:40:29.027169 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-log" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.027188 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-log" Nov 24 14:40:29 crc kubenswrapper[4822]: E1124 14:40:29.027285 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-api" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.027298 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-api" Nov 24 14:40:29 crc kubenswrapper[4822]: E1124 14:40:29.027324 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" containerName="nova-scheduler-scheduler" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.027330 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" containerName="nova-scheduler-scheduler" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.027527 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" containerName="nova-scheduler-scheduler" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.027544 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-log" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.027561 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" containerName="nova-api-api" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.031085 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.034409 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.038887 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.176429 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e55d4de-cf12-4f08-9995-b390b62842b6-logs\") pod \"6e55d4de-cf12-4f08-9995-b390b62842b6\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.176805 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-combined-ca-bundle\") pod \"6e55d4de-cf12-4f08-9995-b390b62842b6\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.176992 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-config-data\") pod \"6e55d4de-cf12-4f08-9995-b390b62842b6\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.177148 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz4jm\" (UniqueName: \"kubernetes.io/projected/6e55d4de-cf12-4f08-9995-b390b62842b6-kube-api-access-wz4jm\") pod \"6e55d4de-cf12-4f08-9995-b390b62842b6\" (UID: \"6e55d4de-cf12-4f08-9995-b390b62842b6\") " Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.177049 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e55d4de-cf12-4f08-9995-b390b62842b6-logs" (OuterVolumeSpecName: "logs") pod "6e55d4de-cf12-4f08-9995-b390b62842b6" (UID: "6e55d4de-cf12-4f08-9995-b390b62842b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.177736 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.177954 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chtqb\" (UniqueName: \"kubernetes.io/projected/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-kube-api-access-chtqb\") pod \"nova-scheduler-0\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.178178 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-config-data\") pod \"nova-scheduler-0\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.178583 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e55d4de-cf12-4f08-9995-b390b62842b6-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.182889 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e55d4de-cf12-4f08-9995-b390b62842b6-kube-api-access-wz4jm" (OuterVolumeSpecName: "kube-api-access-wz4jm") pod "6e55d4de-cf12-4f08-9995-b390b62842b6" (UID: "6e55d4de-cf12-4f08-9995-b390b62842b6"). InnerVolumeSpecName "kube-api-access-wz4jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.207713 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e55d4de-cf12-4f08-9995-b390b62842b6" (UID: "6e55d4de-cf12-4f08-9995-b390b62842b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.217245 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-config-data" (OuterVolumeSpecName: "config-data") pod "6e55d4de-cf12-4f08-9995-b390b62842b6" (UID: "6e55d4de-cf12-4f08-9995-b390b62842b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.280225 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.280322 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chtqb\" (UniqueName: \"kubernetes.io/projected/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-kube-api-access-chtqb\") pod \"nova-scheduler-0\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.280365 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-config-data\") pod \"nova-scheduler-0\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.280448 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.280465 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e55d4de-cf12-4f08-9995-b390b62842b6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.280476 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz4jm\" (UniqueName: \"kubernetes.io/projected/6e55d4de-cf12-4f08-9995-b390b62842b6-kube-api-access-wz4jm\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.283603 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-config-data\") pod \"nova-scheduler-0\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.285485 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.307581 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chtqb\" (UniqueName: \"kubernetes.io/projected/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-kube-api-access-chtqb\") pod \"nova-scheduler-0\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.356762 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.498788 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.499436 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.716727 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c" path="/var/lib/kubelet/pods/ef5e7b8e-fba3-498a-91d1-8554eb0c1c4c/volumes" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.718177 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.891756 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:40:29 crc kubenswrapper[4822]: W1124 14:40:29.896442 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12c70ef7_2d2a_4251_9ad7_1d4bd631c399.slice/crio-a36952c61cb0cdd9a60a20ec775c391ba9b96ea9fe6ae4de0b3db72e63f6f2b3 WatchSource:0}: Error finding container a36952c61cb0cdd9a60a20ec775c391ba9b96ea9fe6ae4de0b3db72e63f6f2b3: Status 404 returned error can't find the container with id a36952c61cb0cdd9a60a20ec775c391ba9b96ea9fe6ae4de0b3db72e63f6f2b3 Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.928741 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"12c70ef7-2d2a-4251-9ad7-1d4bd631c399","Type":"ContainerStarted","Data":"a36952c61cb0cdd9a60a20ec775c391ba9b96ea9fe6ae4de0b3db72e63f6f2b3"} Nov 24 14:40:29 crc kubenswrapper[4822]: I1124 14:40:29.932471 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.006285 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.018649 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.035465 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.038486 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.044363 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.052764 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.202850 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.202932 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22fd7dc3-2bb8-470a-ad82-ade788f78791-logs\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.203022 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2922w\" (UniqueName: \"kubernetes.io/projected/22fd7dc3-2bb8-470a-ad82-ade788f78791-kube-api-access-2922w\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.203076 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-config-data\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.305732 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.305838 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22fd7dc3-2bb8-470a-ad82-ade788f78791-logs\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.305956 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2922w\" (UniqueName: \"kubernetes.io/projected/22fd7dc3-2bb8-470a-ad82-ade788f78791-kube-api-access-2922w\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.306024 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-config-data\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.306469 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22fd7dc3-2bb8-470a-ad82-ade788f78791-logs\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.317440 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-config-data\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.328327 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2922w\" (UniqueName: \"kubernetes.io/projected/22fd7dc3-2bb8-470a-ad82-ade788f78791-kube-api-access-2922w\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.328360 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.397826 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.915739 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.945371 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22fd7dc3-2bb8-470a-ad82-ade788f78791","Type":"ContainerStarted","Data":"0e923707c80f7624aadcc166b1338d9755459ea18ba3c36f56897e9009a2259d"} Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.947491 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"12c70ef7-2d2a-4251-9ad7-1d4bd631c399","Type":"ContainerStarted","Data":"ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe"} Nov 24 14:40:30 crc kubenswrapper[4822]: I1124 14:40:30.977857 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.977837117 podStartE2EDuration="2.977837117s" podCreationTimestamp="2025-11-24 14:40:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:30.972379095 +0000 UTC m=+1268.089019582" watchObservedRunningTime="2025-11-24 14:40:30.977837117 +0000 UTC m=+1268.094477594" Nov 24 14:40:31 crc kubenswrapper[4822]: I1124 14:40:31.720809 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e55d4de-cf12-4f08-9995-b390b62842b6" path="/var/lib/kubelet/pods/6e55d4de-cf12-4f08-9995-b390b62842b6/volumes" Nov 24 14:40:31 crc kubenswrapper[4822]: I1124 14:40:31.974767 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22fd7dc3-2bb8-470a-ad82-ade788f78791","Type":"ContainerStarted","Data":"4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f"} Nov 24 14:40:31 crc kubenswrapper[4822]: I1124 14:40:31.975086 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22fd7dc3-2bb8-470a-ad82-ade788f78791","Type":"ContainerStarted","Data":"e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e"} Nov 24 14:40:31 crc kubenswrapper[4822]: I1124 14:40:31.992229 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.9921903159999999 podStartE2EDuration="1.992190316s" podCreationTimestamp="2025-11-24 14:40:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:31.990910055 +0000 UTC m=+1269.107550552" watchObservedRunningTime="2025-11-24 14:40:31.992190316 +0000 UTC m=+1269.108830783" Nov 24 14:40:33 crc kubenswrapper[4822]: I1124 14:40:33.398109 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:40:33 crc kubenswrapper[4822]: I1124 14:40:33.398416 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e5326828-6e3f-464a-90d6-0cf579f5c598" containerName="kube-state-metrics" containerID="cri-o://7112ed58314143939d290e79da2a1799438f73664f74b747fe08285adc029d52" gracePeriod=30 Nov 24 14:40:33 crc kubenswrapper[4822]: I1124 14:40:33.994813 4822 generic.go:334] "Generic (PLEG): container finished" podID="e5326828-6e3f-464a-90d6-0cf579f5c598" containerID="7112ed58314143939d290e79da2a1799438f73664f74b747fe08285adc029d52" exitCode=2 Nov 24 14:40:33 crc kubenswrapper[4822]: I1124 14:40:33.994878 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e5326828-6e3f-464a-90d6-0cf579f5c598","Type":"ContainerDied","Data":"7112ed58314143939d290e79da2a1799438f73664f74b747fe08285adc029d52"} Nov 24 14:40:33 crc kubenswrapper[4822]: I1124 14:40:33.995112 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e5326828-6e3f-464a-90d6-0cf579f5c598","Type":"ContainerDied","Data":"dc86a6d808d6cb9bc23fdca6be620b51b537cc2d7d07a3e44eb2b01670582c3c"} Nov 24 14:40:33 crc kubenswrapper[4822]: I1124 14:40:33.995129 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc86a6d808d6cb9bc23fdca6be620b51b537cc2d7d07a3e44eb2b01670582c3c" Nov 24 14:40:34 crc kubenswrapper[4822]: I1124 14:40:34.007111 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:40:34 crc kubenswrapper[4822]: I1124 14:40:34.202744 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7vzx\" (UniqueName: \"kubernetes.io/projected/e5326828-6e3f-464a-90d6-0cf579f5c598-kube-api-access-z7vzx\") pod \"e5326828-6e3f-464a-90d6-0cf579f5c598\" (UID: \"e5326828-6e3f-464a-90d6-0cf579f5c598\") " Nov 24 14:40:34 crc kubenswrapper[4822]: I1124 14:40:34.211577 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5326828-6e3f-464a-90d6-0cf579f5c598-kube-api-access-z7vzx" (OuterVolumeSpecName: "kube-api-access-z7vzx") pod "e5326828-6e3f-464a-90d6-0cf579f5c598" (UID: "e5326828-6e3f-464a-90d6-0cf579f5c598"). InnerVolumeSpecName "kube-api-access-z7vzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:34 crc kubenswrapper[4822]: I1124 14:40:34.272107 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 14:40:34 crc kubenswrapper[4822]: I1124 14:40:34.306112 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7vzx\" (UniqueName: \"kubernetes.io/projected/e5326828-6e3f-464a-90d6-0cf579f5c598-kube-api-access-z7vzx\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:34 crc kubenswrapper[4822]: I1124 14:40:34.357302 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:40:34 crc kubenswrapper[4822]: I1124 14:40:34.499596 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:40:34 crc kubenswrapper[4822]: I1124 14:40:34.499635 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.002608 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.035624 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.049216 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.068079 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:40:35 crc kubenswrapper[4822]: E1124 14:40:35.072740 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5326828-6e3f-464a-90d6-0cf579f5c598" containerName="kube-state-metrics" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.072774 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5326828-6e3f-464a-90d6-0cf579f5c598" containerName="kube-state-metrics" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.073003 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5326828-6e3f-464a-90d6-0cf579f5c598" containerName="kube-state-metrics" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.073853 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.075949 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.076165 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.084478 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.225147 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3f75805d-90a9-4782-b65f-8ca7d1b91338-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.225219 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxvvk\" (UniqueName: \"kubernetes.io/projected/3f75805d-90a9-4782-b65f-8ca7d1b91338-kube-api-access-hxvvk\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.225242 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f75805d-90a9-4782-b65f-8ca7d1b91338-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.225346 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f75805d-90a9-4782-b65f-8ca7d1b91338-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.327140 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f75805d-90a9-4782-b65f-8ca7d1b91338-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.327485 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3f75805d-90a9-4782-b65f-8ca7d1b91338-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.327606 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxvvk\" (UniqueName: \"kubernetes.io/projected/3f75805d-90a9-4782-b65f-8ca7d1b91338-kube-api-access-hxvvk\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.327681 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f75805d-90a9-4782-b65f-8ca7d1b91338-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.335912 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f75805d-90a9-4782-b65f-8ca7d1b91338-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.336017 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f75805d-90a9-4782-b65f-8ca7d1b91338-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.336505 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3f75805d-90a9-4782-b65f-8ca7d1b91338-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.360301 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxvvk\" (UniqueName: \"kubernetes.io/projected/3f75805d-90a9-4782-b65f-8ca7d1b91338-kube-api-access-hxvvk\") pod \"kube-state-metrics-0\" (UID: \"3f75805d-90a9-4782-b65f-8ca7d1b91338\") " pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.391021 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.514421 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.216:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.514805 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.216:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.583146 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.583668 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="ceilometer-central-agent" containerID="cri-o://802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037" gracePeriod=30 Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.583876 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="ceilometer-notification-agent" containerID="cri-o://9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80" gracePeriod=30 Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.583903 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="proxy-httpd" containerID="cri-o://cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884" gracePeriod=30 Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.583894 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="sg-core" containerID="cri-o://770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90" gracePeriod=30 Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.735003 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5326828-6e3f-464a-90d6-0cf579f5c598" path="/var/lib/kubelet/pods/e5326828-6e3f-464a-90d6-0cf579f5c598/volumes" Nov 24 14:40:35 crc kubenswrapper[4822]: I1124 14:40:35.865849 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:40:35 crc kubenswrapper[4822]: W1124 14:40:35.868465 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f75805d_90a9_4782_b65f_8ca7d1b91338.slice/crio-86373d9fdf8eee54984ab9ef51159ae9ee2099d85d767312018d11d6cd692d60 WatchSource:0}: Error finding container 86373d9fdf8eee54984ab9ef51159ae9ee2099d85d767312018d11d6cd692d60: Status 404 returned error can't find the container with id 86373d9fdf8eee54984ab9ef51159ae9ee2099d85d767312018d11d6cd692d60 Nov 24 14:40:36 crc kubenswrapper[4822]: I1124 14:40:36.019787 4822 generic.go:334] "Generic (PLEG): container finished" podID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerID="cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884" exitCode=0 Nov 24 14:40:36 crc kubenswrapper[4822]: I1124 14:40:36.019828 4822 generic.go:334] "Generic (PLEG): container finished" podID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerID="770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90" exitCode=2 Nov 24 14:40:36 crc kubenswrapper[4822]: I1124 14:40:36.019837 4822 generic.go:334] "Generic (PLEG): container finished" podID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerID="802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037" exitCode=0 Nov 24 14:40:36 crc kubenswrapper[4822]: I1124 14:40:36.019882 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerDied","Data":"cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884"} Nov 24 14:40:36 crc kubenswrapper[4822]: I1124 14:40:36.019926 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerDied","Data":"770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90"} Nov 24 14:40:36 crc kubenswrapper[4822]: I1124 14:40:36.019939 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerDied","Data":"802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037"} Nov 24 14:40:36 crc kubenswrapper[4822]: I1124 14:40:36.028864 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3f75805d-90a9-4782-b65f-8ca7d1b91338","Type":"ContainerStarted","Data":"86373d9fdf8eee54984ab9ef51159ae9ee2099d85d767312018d11d6cd692d60"} Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.043634 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3f75805d-90a9-4782-b65f-8ca7d1b91338","Type":"ContainerStarted","Data":"c68e9b2b8a8fe191269f5eb8ceaf3533ce5ae14e8ba25222de12508ca87ea704"} Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.045229 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.061301 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.622841823 podStartE2EDuration="2.061282609s" podCreationTimestamp="2025-11-24 14:40:35 +0000 UTC" firstStartedPulling="2025-11-24 14:40:35.870659675 +0000 UTC m=+1272.987300142" lastFinishedPulling="2025-11-24 14:40:36.309100451 +0000 UTC m=+1273.425740928" observedRunningTime="2025-11-24 14:40:37.05907843 +0000 UTC m=+1274.175718927" watchObservedRunningTime="2025-11-24 14:40:37.061282609 +0000 UTC m=+1274.177923086" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.685607 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.779101 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-scripts\") pod \"31d97142-bc40-4398-b4c1-fe12a14f7398\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.779250 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-run-httpd\") pod \"31d97142-bc40-4398-b4c1-fe12a14f7398\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.779340 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-log-httpd\") pod \"31d97142-bc40-4398-b4c1-fe12a14f7398\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.779462 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-sg-core-conf-yaml\") pod \"31d97142-bc40-4398-b4c1-fe12a14f7398\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.779626 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-combined-ca-bundle\") pod \"31d97142-bc40-4398-b4c1-fe12a14f7398\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.779675 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqh98\" (UniqueName: \"kubernetes.io/projected/31d97142-bc40-4398-b4c1-fe12a14f7398-kube-api-access-cqh98\") pod \"31d97142-bc40-4398-b4c1-fe12a14f7398\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.779704 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-config-data\") pod \"31d97142-bc40-4398-b4c1-fe12a14f7398\" (UID: \"31d97142-bc40-4398-b4c1-fe12a14f7398\") " Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.781341 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "31d97142-bc40-4398-b4c1-fe12a14f7398" (UID: "31d97142-bc40-4398-b4c1-fe12a14f7398"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.781798 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "31d97142-bc40-4398-b4c1-fe12a14f7398" (UID: "31d97142-bc40-4398-b4c1-fe12a14f7398"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.791172 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d97142-bc40-4398-b4c1-fe12a14f7398-kube-api-access-cqh98" (OuterVolumeSpecName: "kube-api-access-cqh98") pod "31d97142-bc40-4398-b4c1-fe12a14f7398" (UID: "31d97142-bc40-4398-b4c1-fe12a14f7398"). InnerVolumeSpecName "kube-api-access-cqh98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.795618 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-scripts" (OuterVolumeSpecName: "scripts") pod "31d97142-bc40-4398-b4c1-fe12a14f7398" (UID: "31d97142-bc40-4398-b4c1-fe12a14f7398"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.823905 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "31d97142-bc40-4398-b4c1-fe12a14f7398" (UID: "31d97142-bc40-4398-b4c1-fe12a14f7398"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.873971 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-config-data" (OuterVolumeSpecName: "config-data") pod "31d97142-bc40-4398-b4c1-fe12a14f7398" (UID: "31d97142-bc40-4398-b4c1-fe12a14f7398"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.882933 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.882964 4822 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.882972 4822 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31d97142-bc40-4398-b4c1-fe12a14f7398-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.882981 4822 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.882992 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqh98\" (UniqueName: \"kubernetes.io/projected/31d97142-bc40-4398-b4c1-fe12a14f7398-kube-api-access-cqh98\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.883001 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.916134 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31d97142-bc40-4398-b4c1-fe12a14f7398" (UID: "31d97142-bc40-4398-b4c1-fe12a14f7398"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:37 crc kubenswrapper[4822]: I1124 14:40:37.985571 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d97142-bc40-4398-b4c1-fe12a14f7398-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.062500 4822 generic.go:334] "Generic (PLEG): container finished" podID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerID="9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80" exitCode=0 Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.062666 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerDied","Data":"9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80"} Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.062983 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31d97142-bc40-4398-b4c1-fe12a14f7398","Type":"ContainerDied","Data":"d1cbb61f4f161385b19ceee63778d5d313495e32301479c0e70ce98d18f13fb0"} Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.063009 4822 scope.go:117] "RemoveContainer" containerID="cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.062764 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.128510 4822 scope.go:117] "RemoveContainer" containerID="770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.129225 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.150729 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.158907 4822 scope.go:117] "RemoveContainer" containerID="9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.163194 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:38 crc kubenswrapper[4822]: E1124 14:40:38.163843 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="ceilometer-central-agent" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.163875 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="ceilometer-central-agent" Nov 24 14:40:38 crc kubenswrapper[4822]: E1124 14:40:38.163911 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="sg-core" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.163924 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="sg-core" Nov 24 14:40:38 crc kubenswrapper[4822]: E1124 14:40:38.163964 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="ceilometer-notification-agent" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.163977 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="ceilometer-notification-agent" Nov 24 14:40:38 crc kubenswrapper[4822]: E1124 14:40:38.164034 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="proxy-httpd" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.164045 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="proxy-httpd" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.164471 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="ceilometer-central-agent" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.164518 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="sg-core" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.164538 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="ceilometer-notification-agent" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.164561 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" containerName="proxy-httpd" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.168032 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.171017 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.171110 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.171849 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.172173 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.210935 4822 scope.go:117] "RemoveContainer" containerID="802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.234269 4822 scope.go:117] "RemoveContainer" containerID="cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884" Nov 24 14:40:38 crc kubenswrapper[4822]: E1124 14:40:38.234719 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884\": container with ID starting with cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884 not found: ID does not exist" containerID="cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.234754 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884"} err="failed to get container status \"cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884\": rpc error: code = NotFound desc = could not find container \"cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884\": container with ID starting with cf184582df11be93d5fc086cd255620f0c6845a05f1b25cba5d90b9ebe10f884 not found: ID does not exist" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.234781 4822 scope.go:117] "RemoveContainer" containerID="770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90" Nov 24 14:40:38 crc kubenswrapper[4822]: E1124 14:40:38.235491 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90\": container with ID starting with 770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90 not found: ID does not exist" containerID="770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.235546 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90"} err="failed to get container status \"770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90\": rpc error: code = NotFound desc = could not find container \"770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90\": container with ID starting with 770371d30fe8c97e47c0b9c8afd7551f40363fd1f7955a679f16d14b88be1d90 not found: ID does not exist" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.235580 4822 scope.go:117] "RemoveContainer" containerID="9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80" Nov 24 14:40:38 crc kubenswrapper[4822]: E1124 14:40:38.235905 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80\": container with ID starting with 9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80 not found: ID does not exist" containerID="9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.235934 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80"} err="failed to get container status \"9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80\": rpc error: code = NotFound desc = could not find container \"9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80\": container with ID starting with 9e1548ddb4c85ee995385e3aa6473b22e25c4d7302a8df2aec9e762db5689a80 not found: ID does not exist" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.235955 4822 scope.go:117] "RemoveContainer" containerID="802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037" Nov 24 14:40:38 crc kubenswrapper[4822]: E1124 14:40:38.236240 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037\": container with ID starting with 802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037 not found: ID does not exist" containerID="802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.236268 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037"} err="failed to get container status \"802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037\": rpc error: code = NotFound desc = could not find container \"802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037\": container with ID starting with 802cde70c96e3bb710499356e3f786fc78f2e480fa724557ac166e97460ff037 not found: ID does not exist" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.292181 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-scripts\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.292248 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-run-httpd\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.292313 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-log-httpd\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.292643 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.292711 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-config-data\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.292770 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg974\" (UniqueName: \"kubernetes.io/projected/83dbf687-412c-47ba-8533-688fd2dc6a1d-kube-api-access-mg974\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.292855 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.292984 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.395186 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-scripts\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.395267 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-run-httpd\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.395336 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-log-httpd\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.395458 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.395491 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-config-data\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.395522 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg974\" (UniqueName: \"kubernetes.io/projected/83dbf687-412c-47ba-8533-688fd2dc6a1d-kube-api-access-mg974\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.395558 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.395617 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.395946 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-log-httpd\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.396127 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-run-httpd\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.400934 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-config-data\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.401542 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-scripts\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.407133 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.407298 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.411120 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.417950 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg974\" (UniqueName: \"kubernetes.io/projected/83dbf687-412c-47ba-8533-688fd2dc6a1d-kube-api-access-mg974\") pod \"ceilometer-0\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " pod="openstack/ceilometer-0" Nov 24 14:40:38 crc kubenswrapper[4822]: I1124 14:40:38.509784 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:40:39 crc kubenswrapper[4822]: I1124 14:40:39.041925 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:39 crc kubenswrapper[4822]: I1124 14:40:39.083628 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerStarted","Data":"fd553c041abd3b3a6d0e972f66bbb32c6c83584aefd4e6a46b02a8da33e28e8d"} Nov 24 14:40:39 crc kubenswrapper[4822]: I1124 14:40:39.357233 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 14:40:39 crc kubenswrapper[4822]: I1124 14:40:39.418050 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 14:40:39 crc kubenswrapper[4822]: I1124 14:40:39.715617 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d97142-bc40-4398-b4c1-fe12a14f7398" path="/var/lib/kubelet/pods/31d97142-bc40-4398-b4c1-fe12a14f7398/volumes" Nov 24 14:40:40 crc kubenswrapper[4822]: I1124 14:40:40.104454 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerStarted","Data":"ed54bf9ba2cb8556e6e58da64d10955d5d8ee390029607516d8b275f5cd4c346"} Nov 24 14:40:40 crc kubenswrapper[4822]: I1124 14:40:40.156426 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 14:40:40 crc kubenswrapper[4822]: I1124 14:40:40.398852 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:40:40 crc kubenswrapper[4822]: I1124 14:40:40.398911 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:40:41 crc kubenswrapper[4822]: I1124 14:40:41.122165 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerStarted","Data":"150e93e327aae913416b5879fd99b2216c87224bb18ee1536a304ba99bb91e94"} Nov 24 14:40:41 crc kubenswrapper[4822]: I1124 14:40:41.481450 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:40:41 crc kubenswrapper[4822]: I1124 14:40:41.481458 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:40:42 crc kubenswrapper[4822]: I1124 14:40:42.140930 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerStarted","Data":"5cdd7312b8f31f4da668e8435e18cb21763c4c4b728a4511363de44b15fcdf53"} Nov 24 14:40:43 crc kubenswrapper[4822]: I1124 14:40:43.155079 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerStarted","Data":"2a2ffcf65260a0c02d185bd40db158f206ffa6799b48260a4fd54f5cfd59d59d"} Nov 24 14:40:43 crc kubenswrapper[4822]: I1124 14:40:43.155544 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:40:43 crc kubenswrapper[4822]: I1124 14:40:43.191246 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.573051625 podStartE2EDuration="5.191184484s" podCreationTimestamp="2025-11-24 14:40:38 +0000 UTC" firstStartedPulling="2025-11-24 14:40:39.042452328 +0000 UTC m=+1276.159092805" lastFinishedPulling="2025-11-24 14:40:42.660585147 +0000 UTC m=+1279.777225664" observedRunningTime="2025-11-24 14:40:43.186996221 +0000 UTC m=+1280.303636778" watchObservedRunningTime="2025-11-24 14:40:43.191184484 +0000 UTC m=+1280.307825001" Nov 24 14:40:43 crc kubenswrapper[4822]: E1124 14:40:43.635394 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="4acc7e6a-472b-468a-b709-183f8b3c2b5b" Nov 24 14:40:44 crc kubenswrapper[4822]: I1124 14:40:44.168754 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:40:44 crc kubenswrapper[4822]: I1124 14:40:44.505567 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:40:44 crc kubenswrapper[4822]: I1124 14:40:44.507688 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:40:44 crc kubenswrapper[4822]: I1124 14:40:44.515100 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:40:45 crc kubenswrapper[4822]: I1124 14:40:45.191357 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:40:45 crc kubenswrapper[4822]: I1124 14:40:45.404124 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.106115 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:40:47 crc kubenswrapper[4822]: E1124 14:40:47.106404 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:40:47 crc kubenswrapper[4822]: E1124 14:40:47.106605 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:40:47 crc kubenswrapper[4822]: E1124 14:40:47.106657 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:42:49.106641815 +0000 UTC m=+1406.223282282 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.203985 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.221033 4822 generic.go:334] "Generic (PLEG): container finished" podID="3ed619d2-3396-4407-bbb8-8f042307c43c" containerID="e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c" exitCode=137 Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.221103 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.221115 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3ed619d2-3396-4407-bbb8-8f042307c43c","Type":"ContainerDied","Data":"e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c"} Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.221165 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3ed619d2-3396-4407-bbb8-8f042307c43c","Type":"ContainerDied","Data":"4c117e4d0974524e58449a9bb92367674bfc5c298e764d79b42f5a68563b3bdd"} Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.221187 4822 scope.go:117] "RemoveContainer" containerID="e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.249255 4822 scope.go:117] "RemoveContainer" containerID="e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c" Nov 24 14:40:47 crc kubenswrapper[4822]: E1124 14:40:47.253880 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c\": container with ID starting with e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c not found: ID does not exist" containerID="e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.253932 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c"} err="failed to get container status \"e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c\": rpc error: code = NotFound desc = could not find container \"e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c\": container with ID starting with e2310bd645d8153711d606b473b0c05c8de8593e83685cb9b480bea5c812714c not found: ID does not exist" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.309596 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-combined-ca-bundle\") pod \"3ed619d2-3396-4407-bbb8-8f042307c43c\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.309995 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn7qb\" (UniqueName: \"kubernetes.io/projected/3ed619d2-3396-4407-bbb8-8f042307c43c-kube-api-access-pn7qb\") pod \"3ed619d2-3396-4407-bbb8-8f042307c43c\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.310159 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-config-data\") pod \"3ed619d2-3396-4407-bbb8-8f042307c43c\" (UID: \"3ed619d2-3396-4407-bbb8-8f042307c43c\") " Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.316139 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ed619d2-3396-4407-bbb8-8f042307c43c-kube-api-access-pn7qb" (OuterVolumeSpecName: "kube-api-access-pn7qb") pod "3ed619d2-3396-4407-bbb8-8f042307c43c" (UID: "3ed619d2-3396-4407-bbb8-8f042307c43c"). InnerVolumeSpecName "kube-api-access-pn7qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.340084 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-config-data" (OuterVolumeSpecName: "config-data") pod "3ed619d2-3396-4407-bbb8-8f042307c43c" (UID: "3ed619d2-3396-4407-bbb8-8f042307c43c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.358359 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ed619d2-3396-4407-bbb8-8f042307c43c" (UID: "3ed619d2-3396-4407-bbb8-8f042307c43c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.413436 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.414458 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed619d2-3396-4407-bbb8-8f042307c43c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.414569 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn7qb\" (UniqueName: \"kubernetes.io/projected/3ed619d2-3396-4407-bbb8-8f042307c43c-kube-api-access-pn7qb\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.574689 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.593858 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.607008 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:40:47 crc kubenswrapper[4822]: E1124 14:40:47.607998 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ed619d2-3396-4407-bbb8-8f042307c43c" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.608051 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ed619d2-3396-4407-bbb8-8f042307c43c" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.608624 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ed619d2-3396-4407-bbb8-8f042307c43c" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.610610 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.614933 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.614962 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.615198 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.620869 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.718529 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ed619d2-3396-4407-bbb8-8f042307c43c" path="/var/lib/kubelet/pods/3ed619d2-3396-4407-bbb8-8f042307c43c/volumes" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.720166 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.720293 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd4mb\" (UniqueName: \"kubernetes.io/projected/225bbbb8-99a0-4128-a39c-dc69b59eda7b-kube-api-access-zd4mb\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.720321 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.720433 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.720491 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.823337 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.823593 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.823715 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.823939 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.823977 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd4mb\" (UniqueName: \"kubernetes.io/projected/225bbbb8-99a0-4128-a39c-dc69b59eda7b-kube-api-access-zd4mb\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.828760 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.828767 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.829083 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.829490 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225bbbb8-99a0-4128-a39c-dc69b59eda7b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.841651 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd4mb\" (UniqueName: \"kubernetes.io/projected/225bbbb8-99a0-4128-a39c-dc69b59eda7b-kube-api-access-zd4mb\") pod \"nova-cell1-novncproxy-0\" (UID: \"225bbbb8-99a0-4128-a39c-dc69b59eda7b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:47 crc kubenswrapper[4822]: I1124 14:40:47.933505 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:48 crc kubenswrapper[4822]: I1124 14:40:48.436062 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:40:49 crc kubenswrapper[4822]: I1124 14:40:49.250321 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"225bbbb8-99a0-4128-a39c-dc69b59eda7b","Type":"ContainerStarted","Data":"abb91014a3edbaf5193ac272cbf0d44c26ceb0abf3211a9cb74e0e6d13f03071"} Nov 24 14:40:49 crc kubenswrapper[4822]: I1124 14:40:49.250712 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"225bbbb8-99a0-4128-a39c-dc69b59eda7b","Type":"ContainerStarted","Data":"c8c85efcac8cb28e212def39a07327bdec58267f8302c27c7bbec4df2d0db53b"} Nov 24 14:40:49 crc kubenswrapper[4822]: I1124 14:40:49.288528 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.288503395 podStartE2EDuration="2.288503395s" podCreationTimestamp="2025-11-24 14:40:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:49.277132284 +0000 UTC m=+1286.393772801" watchObservedRunningTime="2025-11-24 14:40:49.288503395 +0000 UTC m=+1286.405143882" Nov 24 14:40:50 crc kubenswrapper[4822]: I1124 14:40:50.402666 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:40:50 crc kubenswrapper[4822]: I1124 14:40:50.403319 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:40:50 crc kubenswrapper[4822]: I1124 14:40:50.406665 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:40:50 crc kubenswrapper[4822]: I1124 14:40:50.407408 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.275941 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.281534 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.524126 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cbd4487cc-66fhk"] Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.526245 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.544373 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cbd4487cc-66fhk"] Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.623609 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-dns-svc\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.623649 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8s44\" (UniqueName: \"kubernetes.io/projected/02fe13aa-aece-4235-b424-e244d720b40f-kube-api-access-x8s44\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.623946 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-config\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.624145 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-nb\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.624414 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-sb\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.726548 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-config\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.726880 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-nb\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.726970 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-sb\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.727024 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-dns-svc\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.727043 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8s44\" (UniqueName: \"kubernetes.io/projected/02fe13aa-aece-4235-b424-e244d720b40f-kube-api-access-x8s44\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.727792 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-nb\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.727816 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-dns-svc\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.727797 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-config\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.727913 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-sb\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.752111 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8s44\" (UniqueName: \"kubernetes.io/projected/02fe13aa-aece-4235-b424-e244d720b40f-kube-api-access-x8s44\") pod \"dnsmasq-dns-7cbd4487cc-66fhk\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:51 crc kubenswrapper[4822]: I1124 14:40:51.845881 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:52 crc kubenswrapper[4822]: I1124 14:40:52.360350 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cbd4487cc-66fhk"] Nov 24 14:40:52 crc kubenswrapper[4822]: I1124 14:40:52.934548 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:53 crc kubenswrapper[4822]: I1124 14:40:53.294180 4822 generic.go:334] "Generic (PLEG): container finished" podID="02fe13aa-aece-4235-b424-e244d720b40f" containerID="83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1" exitCode=0 Nov 24 14:40:53 crc kubenswrapper[4822]: I1124 14:40:53.294280 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" event={"ID":"02fe13aa-aece-4235-b424-e244d720b40f","Type":"ContainerDied","Data":"83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1"} Nov 24 14:40:53 crc kubenswrapper[4822]: I1124 14:40:53.294361 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" event={"ID":"02fe13aa-aece-4235-b424-e244d720b40f","Type":"ContainerStarted","Data":"0c6ed307f8689199010e8e90d6673870ad734dc501dffb0803ab3ae3b5a19817"} Nov 24 14:40:53 crc kubenswrapper[4822]: I1124 14:40:53.567229 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:53 crc kubenswrapper[4822]: I1124 14:40:53.567984 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="ceilometer-central-agent" containerID="cri-o://ed54bf9ba2cb8556e6e58da64d10955d5d8ee390029607516d8b275f5cd4c346" gracePeriod=30 Nov 24 14:40:53 crc kubenswrapper[4822]: I1124 14:40:53.568061 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="sg-core" containerID="cri-o://5cdd7312b8f31f4da668e8435e18cb21763c4c4b728a4511363de44b15fcdf53" gracePeriod=30 Nov 24 14:40:53 crc kubenswrapper[4822]: I1124 14:40:53.568101 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="proxy-httpd" containerID="cri-o://2a2ffcf65260a0c02d185bd40db158f206ffa6799b48260a4fd54f5cfd59d59d" gracePeriod=30 Nov 24 14:40:53 crc kubenswrapper[4822]: I1124 14:40:53.568155 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="ceilometer-notification-agent" containerID="cri-o://150e93e327aae913416b5879fd99b2216c87224bb18ee1536a304ba99bb91e94" gracePeriod=30 Nov 24 14:40:53 crc kubenswrapper[4822]: I1124 14:40:53.582096 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.220:3000/\": EOF" Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.218606 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.310509 4822 generic.go:334] "Generic (PLEG): container finished" podID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerID="2a2ffcf65260a0c02d185bd40db158f206ffa6799b48260a4fd54f5cfd59d59d" exitCode=0 Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.310811 4822 generic.go:334] "Generic (PLEG): container finished" podID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerID="5cdd7312b8f31f4da668e8435e18cb21763c4c4b728a4511363de44b15fcdf53" exitCode=2 Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.310825 4822 generic.go:334] "Generic (PLEG): container finished" podID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerID="ed54bf9ba2cb8556e6e58da64d10955d5d8ee390029607516d8b275f5cd4c346" exitCode=0 Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.310597 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerDied","Data":"2a2ffcf65260a0c02d185bd40db158f206ffa6799b48260a4fd54f5cfd59d59d"} Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.310910 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerDied","Data":"5cdd7312b8f31f4da668e8435e18cb21763c4c4b728a4511363de44b15fcdf53"} Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.310929 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerDied","Data":"ed54bf9ba2cb8556e6e58da64d10955d5d8ee390029607516d8b275f5cd4c346"} Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.313444 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" event={"ID":"02fe13aa-aece-4235-b424-e244d720b40f","Type":"ContainerStarted","Data":"aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082"} Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.313648 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-log" containerID="cri-o://e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e" gracePeriod=30 Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.314393 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-api" containerID="cri-o://4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f" gracePeriod=30 Nov 24 14:40:54 crc kubenswrapper[4822]: I1124 14:40:54.353869 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" podStartSLOduration=3.353843549 podStartE2EDuration="3.353843549s" podCreationTimestamp="2025-11-24 14:40:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:40:54.344828773 +0000 UTC m=+1291.461469270" watchObservedRunningTime="2025-11-24 14:40:54.353843549 +0000 UTC m=+1291.470484026" Nov 24 14:40:55 crc kubenswrapper[4822]: I1124 14:40:55.327007 4822 generic.go:334] "Generic (PLEG): container finished" podID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerID="e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e" exitCode=143 Nov 24 14:40:55 crc kubenswrapper[4822]: I1124 14:40:55.327085 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22fd7dc3-2bb8-470a-ad82-ade788f78791","Type":"ContainerDied","Data":"e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e"} Nov 24 14:40:55 crc kubenswrapper[4822]: I1124 14:40:55.327283 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.345382 4822 generic.go:334] "Generic (PLEG): container finished" podID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerID="150e93e327aae913416b5879fd99b2216c87224bb18ee1536a304ba99bb91e94" exitCode=0 Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.345450 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerDied","Data":"150e93e327aae913416b5879fd99b2216c87224bb18ee1536a304ba99bb91e94"} Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.345899 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83dbf687-412c-47ba-8533-688fd2dc6a1d","Type":"ContainerDied","Data":"fd553c041abd3b3a6d0e972f66bbb32c6c83584aefd4e6a46b02a8da33e28e8d"} Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.345940 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd553c041abd3b3a6d0e972f66bbb32c6c83584aefd4e6a46b02a8da33e28e8d" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.427152 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.546338 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-combined-ca-bundle\") pod \"83dbf687-412c-47ba-8533-688fd2dc6a1d\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.546400 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-ceilometer-tls-certs\") pod \"83dbf687-412c-47ba-8533-688fd2dc6a1d\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.546463 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-run-httpd\") pod \"83dbf687-412c-47ba-8533-688fd2dc6a1d\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.546585 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-scripts\") pod \"83dbf687-412c-47ba-8533-688fd2dc6a1d\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.546647 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-config-data\") pod \"83dbf687-412c-47ba-8533-688fd2dc6a1d\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.546695 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg974\" (UniqueName: \"kubernetes.io/projected/83dbf687-412c-47ba-8533-688fd2dc6a1d-kube-api-access-mg974\") pod \"83dbf687-412c-47ba-8533-688fd2dc6a1d\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.546723 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-sg-core-conf-yaml\") pod \"83dbf687-412c-47ba-8533-688fd2dc6a1d\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.546776 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-log-httpd\") pod \"83dbf687-412c-47ba-8533-688fd2dc6a1d\" (UID: \"83dbf687-412c-47ba-8533-688fd2dc6a1d\") " Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.547292 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "83dbf687-412c-47ba-8533-688fd2dc6a1d" (UID: "83dbf687-412c-47ba-8533-688fd2dc6a1d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.547935 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "83dbf687-412c-47ba-8533-688fd2dc6a1d" (UID: "83dbf687-412c-47ba-8533-688fd2dc6a1d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.564525 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83dbf687-412c-47ba-8533-688fd2dc6a1d-kube-api-access-mg974" (OuterVolumeSpecName: "kube-api-access-mg974") pod "83dbf687-412c-47ba-8533-688fd2dc6a1d" (UID: "83dbf687-412c-47ba-8533-688fd2dc6a1d"). InnerVolumeSpecName "kube-api-access-mg974". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.565285 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-scripts" (OuterVolumeSpecName: "scripts") pod "83dbf687-412c-47ba-8533-688fd2dc6a1d" (UID: "83dbf687-412c-47ba-8533-688fd2dc6a1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.609361 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "83dbf687-412c-47ba-8533-688fd2dc6a1d" (UID: "83dbf687-412c-47ba-8533-688fd2dc6a1d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.629810 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83dbf687-412c-47ba-8533-688fd2dc6a1d" (UID: "83dbf687-412c-47ba-8533-688fd2dc6a1d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.652885 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.652918 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg974\" (UniqueName: \"kubernetes.io/projected/83dbf687-412c-47ba-8533-688fd2dc6a1d-kube-api-access-mg974\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.652934 4822 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.652946 4822 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.652958 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.652969 4822 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83dbf687-412c-47ba-8533-688fd2dc6a1d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.669377 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "83dbf687-412c-47ba-8533-688fd2dc6a1d" (UID: "83dbf687-412c-47ba-8533-688fd2dc6a1d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.715884 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-config-data" (OuterVolumeSpecName: "config-data") pod "83dbf687-412c-47ba-8533-688fd2dc6a1d" (UID: "83dbf687-412c-47ba-8533-688fd2dc6a1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.754555 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:56 crc kubenswrapper[4822]: I1124 14:40:56.754779 4822 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/83dbf687-412c-47ba-8533-688fd2dc6a1d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.357748 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.418820 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.432080 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.473912 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:57 crc kubenswrapper[4822]: E1124 14:40:57.476565 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="proxy-httpd" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.476632 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="proxy-httpd" Nov 24 14:40:57 crc kubenswrapper[4822]: E1124 14:40:57.476693 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="ceilometer-notification-agent" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.476707 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="ceilometer-notification-agent" Nov 24 14:40:57 crc kubenswrapper[4822]: E1124 14:40:57.476759 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="sg-core" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.476771 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="sg-core" Nov 24 14:40:57 crc kubenswrapper[4822]: E1124 14:40:57.476813 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="ceilometer-central-agent" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.476826 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="ceilometer-central-agent" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.478511 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="proxy-httpd" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.479110 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="sg-core" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.479178 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="ceilometer-notification-agent" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.479219 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" containerName="ceilometer-central-agent" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.483302 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.487307 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.487462 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.487317 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.491356 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.582519 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.582895 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-scripts\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.582943 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8fjd\" (UniqueName: \"kubernetes.io/projected/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-kube-api-access-l8fjd\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.583178 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-config-data\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.583238 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.583302 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-log-httpd\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.583524 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-run-httpd\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.583621 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.685686 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.685856 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.685902 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-scripts\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.685943 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8fjd\" (UniqueName: \"kubernetes.io/projected/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-kube-api-access-l8fjd\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.686125 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-config-data\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.686165 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.686295 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-log-httpd\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.686371 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-run-httpd\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.687182 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-run-httpd\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.687591 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-log-httpd\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.690981 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.691327 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-config-data\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.691472 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.706295 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.706447 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-scripts\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.712615 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8fjd\" (UniqueName: \"kubernetes.io/projected/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-kube-api-access-l8fjd\") pod \"ceilometer-0\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.747921 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83dbf687-412c-47ba-8533-688fd2dc6a1d" path="/var/lib/kubelet/pods/83dbf687-412c-47ba-8533-688fd2dc6a1d/volumes" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.926596 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.934445 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:57 crc kubenswrapper[4822]: I1124 14:40:57.957662 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.039233 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.196078 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-combined-ca-bundle\") pod \"22fd7dc3-2bb8-470a-ad82-ade788f78791\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.196232 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2922w\" (UniqueName: \"kubernetes.io/projected/22fd7dc3-2bb8-470a-ad82-ade788f78791-kube-api-access-2922w\") pod \"22fd7dc3-2bb8-470a-ad82-ade788f78791\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.196312 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-config-data\") pod \"22fd7dc3-2bb8-470a-ad82-ade788f78791\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.196454 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22fd7dc3-2bb8-470a-ad82-ade788f78791-logs\") pod \"22fd7dc3-2bb8-470a-ad82-ade788f78791\" (UID: \"22fd7dc3-2bb8-470a-ad82-ade788f78791\") " Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.197253 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22fd7dc3-2bb8-470a-ad82-ade788f78791-logs" (OuterVolumeSpecName: "logs") pod "22fd7dc3-2bb8-470a-ad82-ade788f78791" (UID: "22fd7dc3-2bb8-470a-ad82-ade788f78791"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.197650 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22fd7dc3-2bb8-470a-ad82-ade788f78791-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.218654 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22fd7dc3-2bb8-470a-ad82-ade788f78791-kube-api-access-2922w" (OuterVolumeSpecName: "kube-api-access-2922w") pod "22fd7dc3-2bb8-470a-ad82-ade788f78791" (UID: "22fd7dc3-2bb8-470a-ad82-ade788f78791"). InnerVolumeSpecName "kube-api-access-2922w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.229336 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-config-data" (OuterVolumeSpecName: "config-data") pod "22fd7dc3-2bb8-470a-ad82-ade788f78791" (UID: "22fd7dc3-2bb8-470a-ad82-ade788f78791"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.231382 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22fd7dc3-2bb8-470a-ad82-ade788f78791" (UID: "22fd7dc3-2bb8-470a-ad82-ade788f78791"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.300269 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.300558 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2922w\" (UniqueName: \"kubernetes.io/projected/22fd7dc3-2bb8-470a-ad82-ade788f78791-kube-api-access-2922w\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.300601 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd7dc3-2bb8-470a-ad82-ade788f78791-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.369785 4822 generic.go:334] "Generic (PLEG): container finished" podID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerID="4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f" exitCode=0 Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.369839 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.369931 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22fd7dc3-2bb8-470a-ad82-ade788f78791","Type":"ContainerDied","Data":"4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f"} Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.369970 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22fd7dc3-2bb8-470a-ad82-ade788f78791","Type":"ContainerDied","Data":"0e923707c80f7624aadcc166b1338d9755459ea18ba3c36f56897e9009a2259d"} Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.369995 4822 scope.go:117] "RemoveContainer" containerID="4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.386338 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.399222 4822 scope.go:117] "RemoveContainer" containerID="e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.444356 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.449146 4822 scope.go:117] "RemoveContainer" containerID="4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f" Nov 24 14:40:58 crc kubenswrapper[4822]: E1124 14:40:58.449991 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f\": container with ID starting with 4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f not found: ID does not exist" containerID="4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.450025 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f"} err="failed to get container status \"4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f\": rpc error: code = NotFound desc = could not find container \"4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f\": container with ID starting with 4cb1a0a2c671b33572b8e834d82ebe4b1bb51e9be791b2470f08c17b4f34a51f not found: ID does not exist" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.450046 4822 scope.go:117] "RemoveContainer" containerID="e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e" Nov 24 14:40:58 crc kubenswrapper[4822]: E1124 14:40:58.455781 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e\": container with ID starting with e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e not found: ID does not exist" containerID="e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.455837 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e"} err="failed to get container status \"e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e\": rpc error: code = NotFound desc = could not find container \"e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e\": container with ID starting with e3286319c8751f423e47341b40240d641bf6c712eca272fca8012b7f0deb0c5e not found: ID does not exist" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.461604 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.481111 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:58 crc kubenswrapper[4822]: E1124 14:40:58.481545 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-log" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.481564 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-log" Nov 24 14:40:58 crc kubenswrapper[4822]: E1124 14:40:58.481579 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-api" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.481586 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-api" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.481817 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-api" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.481840 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" containerName="nova-api-log" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.482979 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.487168 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.487315 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.489711 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.493190 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.502778 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.608894 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-public-tls-certs\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.609003 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtnn2\" (UniqueName: \"kubernetes.io/projected/dd9412c3-1984-48c1-9fa6-50cde5997fc7-kube-api-access-mtnn2\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.609025 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd9412c3-1984-48c1-9fa6-50cde5997fc7-logs\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.609054 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.609124 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.609162 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-config-data\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.657434 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-sxqwl"] Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.659063 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.661542 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.661817 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.671340 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-sxqwl"] Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.710581 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-public-tls-certs\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.710709 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtnn2\" (UniqueName: \"kubernetes.io/projected/dd9412c3-1984-48c1-9fa6-50cde5997fc7-kube-api-access-mtnn2\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.710731 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd9412c3-1984-48c1-9fa6-50cde5997fc7-logs\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.710762 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.710793 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.710829 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-config-data\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.711700 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd9412c3-1984-48c1-9fa6-50cde5997fc7-logs\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.715666 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.715935 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-public-tls-certs\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.716628 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.716695 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-config-data\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.731218 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtnn2\" (UniqueName: \"kubernetes.io/projected/dd9412c3-1984-48c1-9fa6-50cde5997fc7-kube-api-access-mtnn2\") pod \"nova-api-0\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.812976 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.813331 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-scripts\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.813481 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-config-data\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.813837 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvlsz\" (UniqueName: \"kubernetes.io/projected/75e3bf9c-9f8b-4713-ba67-3e21635c4521-kube-api-access-mvlsz\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.845784 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.917878 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.917990 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-scripts\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.918810 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-config-data\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.919081 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvlsz\" (UniqueName: \"kubernetes.io/projected/75e3bf9c-9f8b-4713-ba67-3e21635c4521-kube-api-access-mvlsz\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.922510 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-scripts\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.923453 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.934772 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-config-data\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.943821 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvlsz\" (UniqueName: \"kubernetes.io/projected/75e3bf9c-9f8b-4713-ba67-3e21635c4521-kube-api-access-mvlsz\") pod \"nova-cell1-cell-mapping-sxqwl\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:58 crc kubenswrapper[4822]: I1124 14:40:58.978553 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:40:59 crc kubenswrapper[4822]: I1124 14:40:59.369242 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:40:59 crc kubenswrapper[4822]: I1124 14:40:59.411072 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerStarted","Data":"c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a"} Nov 24 14:40:59 crc kubenswrapper[4822]: I1124 14:40:59.411117 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerStarted","Data":"bce2d94199a0959f5cbe78669d333d7137dfacc85c99c152f6c2dc4b80d60f27"} Nov 24 14:40:59 crc kubenswrapper[4822]: I1124 14:40:59.416839 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd9412c3-1984-48c1-9fa6-50cde5997fc7","Type":"ContainerStarted","Data":"94ebc791d18150f9be2039285ba7d011e9e1b0aa49e1f1e6cfb8921969cf581e"} Nov 24 14:40:59 crc kubenswrapper[4822]: I1124 14:40:59.599889 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-sxqwl"] Nov 24 14:40:59 crc kubenswrapper[4822]: W1124 14:40:59.608019 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75e3bf9c_9f8b_4713_ba67_3e21635c4521.slice/crio-968124a4481626ea657fe0b209235f5ff5b805604101b706fff67c019365c1f6 WatchSource:0}: Error finding container 968124a4481626ea657fe0b209235f5ff5b805604101b706fff67c019365c1f6: Status 404 returned error can't find the container with id 968124a4481626ea657fe0b209235f5ff5b805604101b706fff67c019365c1f6 Nov 24 14:40:59 crc kubenswrapper[4822]: I1124 14:40:59.730449 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22fd7dc3-2bb8-470a-ad82-ade788f78791" path="/var/lib/kubelet/pods/22fd7dc3-2bb8-470a-ad82-ade788f78791/volumes" Nov 24 14:41:00 crc kubenswrapper[4822]: I1124 14:41:00.428860 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerStarted","Data":"325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301"} Nov 24 14:41:00 crc kubenswrapper[4822]: I1124 14:41:00.433763 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sxqwl" event={"ID":"75e3bf9c-9f8b-4713-ba67-3e21635c4521","Type":"ContainerStarted","Data":"b56c541e746656bb8ec65b3233905aaa6c9db78ea61d3d126e1ef7b3110c3306"} Nov 24 14:41:00 crc kubenswrapper[4822]: I1124 14:41:00.433791 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sxqwl" event={"ID":"75e3bf9c-9f8b-4713-ba67-3e21635c4521","Type":"ContainerStarted","Data":"968124a4481626ea657fe0b209235f5ff5b805604101b706fff67c019365c1f6"} Nov 24 14:41:00 crc kubenswrapper[4822]: I1124 14:41:00.436840 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd9412c3-1984-48c1-9fa6-50cde5997fc7","Type":"ContainerStarted","Data":"281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8"} Nov 24 14:41:00 crc kubenswrapper[4822]: I1124 14:41:00.436899 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd9412c3-1984-48c1-9fa6-50cde5997fc7","Type":"ContainerStarted","Data":"bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39"} Nov 24 14:41:00 crc kubenswrapper[4822]: I1124 14:41:00.459084 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-sxqwl" podStartSLOduration=2.459061837 podStartE2EDuration="2.459061837s" podCreationTimestamp="2025-11-24 14:40:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:00.454009075 +0000 UTC m=+1297.570649552" watchObservedRunningTime="2025-11-24 14:41:00.459061837 +0000 UTC m=+1297.575702314" Nov 24 14:41:00 crc kubenswrapper[4822]: I1124 14:41:00.493076 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.493045216 podStartE2EDuration="2.493045216s" podCreationTimestamp="2025-11-24 14:40:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:00.481306373 +0000 UTC m=+1297.597946860" watchObservedRunningTime="2025-11-24 14:41:00.493045216 +0000 UTC m=+1297.609685693" Nov 24 14:41:01 crc kubenswrapper[4822]: I1124 14:41:01.460089 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerStarted","Data":"4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42"} Nov 24 14:41:01 crc kubenswrapper[4822]: I1124 14:41:01.848259 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:41:01 crc kubenswrapper[4822]: I1124 14:41:01.973092 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b97cb559f-944lv"] Nov 24 14:41:01 crc kubenswrapper[4822]: I1124 14:41:01.973531 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" podUID="7ac64487-4fa7-42aa-a026-feedaaff7a94" containerName="dnsmasq-dns" containerID="cri-o://bb7d3198e609a9fd1f2ea6654d223909ccd890bbabc87a7ba4e1e7cf06c07825" gracePeriod=10 Nov 24 14:41:02 crc kubenswrapper[4822]: E1124 14:41:02.321713 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" podUID="753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.485100 4822 generic.go:334] "Generic (PLEG): container finished" podID="7ac64487-4fa7-42aa-a026-feedaaff7a94" containerID="bb7d3198e609a9fd1f2ea6654d223909ccd890bbabc87a7ba4e1e7cf06c07825" exitCode=0 Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.485183 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" event={"ID":"7ac64487-4fa7-42aa-a026-feedaaff7a94","Type":"ContainerDied","Data":"bb7d3198e609a9fd1f2ea6654d223909ccd890bbabc87a7ba4e1e7cf06c07825"} Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.495224 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.496686 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerStarted","Data":"df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe"} Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.496737 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.526428 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.209901352 podStartE2EDuration="5.526404508s" podCreationTimestamp="2025-11-24 14:40:57 +0000 UTC" firstStartedPulling="2025-11-24 14:40:58.491816925 +0000 UTC m=+1295.608457402" lastFinishedPulling="2025-11-24 14:41:01.808320061 +0000 UTC m=+1298.924960558" observedRunningTime="2025-11-24 14:41:02.523158445 +0000 UTC m=+1299.639798942" watchObservedRunningTime="2025-11-24 14:41:02.526404508 +0000 UTC m=+1299.643044985" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.547767 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.728182 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-sb\") pod \"7ac64487-4fa7-42aa-a026-feedaaff7a94\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.728335 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-nb\") pod \"7ac64487-4fa7-42aa-a026-feedaaff7a94\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.728421 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-config\") pod \"7ac64487-4fa7-42aa-a026-feedaaff7a94\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.729153 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-dns-svc\") pod \"7ac64487-4fa7-42aa-a026-feedaaff7a94\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.729265 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmdr6\" (UniqueName: \"kubernetes.io/projected/7ac64487-4fa7-42aa-a026-feedaaff7a94-kube-api-access-vmdr6\") pod \"7ac64487-4fa7-42aa-a026-feedaaff7a94\" (UID: \"7ac64487-4fa7-42aa-a026-feedaaff7a94\") " Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.735521 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac64487-4fa7-42aa-a026-feedaaff7a94-kube-api-access-vmdr6" (OuterVolumeSpecName: "kube-api-access-vmdr6") pod "7ac64487-4fa7-42aa-a026-feedaaff7a94" (UID: "7ac64487-4fa7-42aa-a026-feedaaff7a94"). InnerVolumeSpecName "kube-api-access-vmdr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.776224 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-config" (OuterVolumeSpecName: "config") pod "7ac64487-4fa7-42aa-a026-feedaaff7a94" (UID: "7ac64487-4fa7-42aa-a026-feedaaff7a94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.788240 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7ac64487-4fa7-42aa-a026-feedaaff7a94" (UID: "7ac64487-4fa7-42aa-a026-feedaaff7a94"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.788676 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7ac64487-4fa7-42aa-a026-feedaaff7a94" (UID: "7ac64487-4fa7-42aa-a026-feedaaff7a94"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.789560 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7ac64487-4fa7-42aa-a026-feedaaff7a94" (UID: "7ac64487-4fa7-42aa-a026-feedaaff7a94"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.831294 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.831329 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.831338 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.831349 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ac64487-4fa7-42aa-a026-feedaaff7a94-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:02 crc kubenswrapper[4822]: I1124 14:41:02.831358 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmdr6\" (UniqueName: \"kubernetes.io/projected/7ac64487-4fa7-42aa-a026-feedaaff7a94-kube-api-access-vmdr6\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:03 crc kubenswrapper[4822]: I1124 14:41:03.522995 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" event={"ID":"7ac64487-4fa7-42aa-a026-feedaaff7a94","Type":"ContainerDied","Data":"56ed15770927c4aa78df8320489697f13901a1c930d5853605f81cdbcae48136"} Nov 24 14:41:03 crc kubenswrapper[4822]: I1124 14:41:03.523369 4822 scope.go:117] "RemoveContainer" containerID="bb7d3198e609a9fd1f2ea6654d223909ccd890bbabc87a7ba4e1e7cf06c07825" Nov 24 14:41:03 crc kubenswrapper[4822]: I1124 14:41:03.523050 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b97cb559f-944lv" Nov 24 14:41:03 crc kubenswrapper[4822]: I1124 14:41:03.564647 4822 scope.go:117] "RemoveContainer" containerID="3cff95b80e493ae40a3886e8bef4b414e4e8b5154aad25f11b406bd789437130" Nov 24 14:41:03 crc kubenswrapper[4822]: I1124 14:41:03.611822 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b97cb559f-944lv"] Nov 24 14:41:03 crc kubenswrapper[4822]: I1124 14:41:03.618404 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b97cb559f-944lv"] Nov 24 14:41:03 crc kubenswrapper[4822]: I1124 14:41:03.715884 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac64487-4fa7-42aa-a026-feedaaff7a94" path="/var/lib/kubelet/pods/7ac64487-4fa7-42aa-a026-feedaaff7a94/volumes" Nov 24 14:41:04 crc kubenswrapper[4822]: I1124 14:41:04.536838 4822 generic.go:334] "Generic (PLEG): container finished" podID="75e3bf9c-9f8b-4713-ba67-3e21635c4521" containerID="b56c541e746656bb8ec65b3233905aaa6c9db78ea61d3d126e1ef7b3110c3306" exitCode=0 Nov 24 14:41:04 crc kubenswrapper[4822]: I1124 14:41:04.536894 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sxqwl" event={"ID":"75e3bf9c-9f8b-4713-ba67-3e21635c4521","Type":"ContainerDied","Data":"b56c541e746656bb8ec65b3233905aaa6c9db78ea61d3d126e1ef7b3110c3306"} Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.041992 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.213253 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-scripts\") pod \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.213365 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvlsz\" (UniqueName: \"kubernetes.io/projected/75e3bf9c-9f8b-4713-ba67-3e21635c4521-kube-api-access-mvlsz\") pod \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.213418 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-combined-ca-bundle\") pod \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.213660 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-config-data\") pod \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\" (UID: \"75e3bf9c-9f8b-4713-ba67-3e21635c4521\") " Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.220267 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-scripts" (OuterVolumeSpecName: "scripts") pod "75e3bf9c-9f8b-4713-ba67-3e21635c4521" (UID: "75e3bf9c-9f8b-4713-ba67-3e21635c4521"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.221067 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75e3bf9c-9f8b-4713-ba67-3e21635c4521-kube-api-access-mvlsz" (OuterVolumeSpecName: "kube-api-access-mvlsz") pod "75e3bf9c-9f8b-4713-ba67-3e21635c4521" (UID: "75e3bf9c-9f8b-4713-ba67-3e21635c4521"). InnerVolumeSpecName "kube-api-access-mvlsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.259539 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-config-data" (OuterVolumeSpecName: "config-data") pod "75e3bf9c-9f8b-4713-ba67-3e21635c4521" (UID: "75e3bf9c-9f8b-4713-ba67-3e21635c4521"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.276047 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75e3bf9c-9f8b-4713-ba67-3e21635c4521" (UID: "75e3bf9c-9f8b-4713-ba67-3e21635c4521"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.318117 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvlsz\" (UniqueName: \"kubernetes.io/projected/75e3bf9c-9f8b-4713-ba67-3e21635c4521-kube-api-access-mvlsz\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.318167 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.318185 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.318228 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e3bf9c-9f8b-4713-ba67-3e21635c4521-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.569441 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sxqwl" event={"ID":"75e3bf9c-9f8b-4713-ba67-3e21635c4521","Type":"ContainerDied","Data":"968124a4481626ea657fe0b209235f5ff5b805604101b706fff67c019365c1f6"} Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.569508 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="968124a4481626ea657fe0b209235f5ff5b805604101b706fff67c019365c1f6" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.570399 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sxqwl" Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.795633 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.796154 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerName="nova-api-log" containerID="cri-o://bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39" gracePeriod=30 Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.796291 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerName="nova-api-api" containerID="cri-o://281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8" gracePeriod=30 Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.807836 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.808062 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="12c70ef7-2d2a-4251-9ad7-1d4bd631c399" containerName="nova-scheduler-scheduler" containerID="cri-o://ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe" gracePeriod=30 Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.840632 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.840968 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-log" containerID="cri-o://63b708115911fa584ad96a7bf285d6763f12a71f178e4989618107f79a33305d" gracePeriod=30 Nov 24 14:41:06 crc kubenswrapper[4822]: I1124 14:41:06.841337 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-metadata" containerID="cri-o://e6533dec0b4a77a1cc2e167c2758c732b27d3561356b0dd94e42324fb97a8c32" gracePeriod=30 Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.446362 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.446512 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.446710 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.446769 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:43:09.446750027 +0000 UTC m=+1426.563390504 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.454784 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.547543 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-config-data\") pod \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.547593 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-combined-ca-bundle\") pod \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.547614 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-public-tls-certs\") pod \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.547668 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtnn2\" (UniqueName: \"kubernetes.io/projected/dd9412c3-1984-48c1-9fa6-50cde5997fc7-kube-api-access-mtnn2\") pod \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.547778 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-internal-tls-certs\") pod \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.547849 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd9412c3-1984-48c1-9fa6-50cde5997fc7-logs\") pod \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\" (UID: \"dd9412c3-1984-48c1-9fa6-50cde5997fc7\") " Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.548514 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd9412c3-1984-48c1-9fa6-50cde5997fc7-logs" (OuterVolumeSpecName: "logs") pod "dd9412c3-1984-48c1-9fa6-50cde5997fc7" (UID: "dd9412c3-1984-48c1-9fa6-50cde5997fc7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.558869 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd9412c3-1984-48c1-9fa6-50cde5997fc7-kube-api-access-mtnn2" (OuterVolumeSpecName: "kube-api-access-mtnn2") pod "dd9412c3-1984-48c1-9fa6-50cde5997fc7" (UID: "dd9412c3-1984-48c1-9fa6-50cde5997fc7"). InnerVolumeSpecName "kube-api-access-mtnn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.587624 4822 generic.go:334] "Generic (PLEG): container finished" podID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerID="63b708115911fa584ad96a7bf285d6763f12a71f178e4989618107f79a33305d" exitCode=143 Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.587699 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0962a558-6bb1-44fe-953a-1e4f717b175e","Type":"ContainerDied","Data":"63b708115911fa584ad96a7bf285d6763f12a71f178e4989618107f79a33305d"} Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.589540 4822 generic.go:334] "Generic (PLEG): container finished" podID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerID="281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8" exitCode=0 Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.589565 4822 generic.go:334] "Generic (PLEG): container finished" podID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerID="bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39" exitCode=143 Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.589583 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd9412c3-1984-48c1-9fa6-50cde5997fc7","Type":"ContainerDied","Data":"281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8"} Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.589605 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd9412c3-1984-48c1-9fa6-50cde5997fc7","Type":"ContainerDied","Data":"bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39"} Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.589617 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd9412c3-1984-48c1-9fa6-50cde5997fc7","Type":"ContainerDied","Data":"94ebc791d18150f9be2039285ba7d011e9e1b0aa49e1f1e6cfb8921969cf581e"} Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.589638 4822 scope.go:117] "RemoveContainer" containerID="281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.589780 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.603147 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-config-data" (OuterVolumeSpecName: "config-data") pod "dd9412c3-1984-48c1-9fa6-50cde5997fc7" (UID: "dd9412c3-1984-48c1-9fa6-50cde5997fc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.609547 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd9412c3-1984-48c1-9fa6-50cde5997fc7" (UID: "dd9412c3-1984-48c1-9fa6-50cde5997fc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.633427 4822 scope.go:117] "RemoveContainer" containerID="bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.644931 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "dd9412c3-1984-48c1-9fa6-50cde5997fc7" (UID: "dd9412c3-1984-48c1-9fa6-50cde5997fc7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.650879 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtnn2\" (UniqueName: \"kubernetes.io/projected/dd9412c3-1984-48c1-9fa6-50cde5997fc7-kube-api-access-mtnn2\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.650906 4822 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.650933 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd9412c3-1984-48c1-9fa6-50cde5997fc7-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.650946 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.650954 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.656305 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dd9412c3-1984-48c1-9fa6-50cde5997fc7" (UID: "dd9412c3-1984-48c1-9fa6-50cde5997fc7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.676688 4822 scope.go:117] "RemoveContainer" containerID="281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8" Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.677266 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8\": container with ID starting with 281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8 not found: ID does not exist" containerID="281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.677298 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8"} err="failed to get container status \"281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8\": rpc error: code = NotFound desc = could not find container \"281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8\": container with ID starting with 281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8 not found: ID does not exist" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.677317 4822 scope.go:117] "RemoveContainer" containerID="bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39" Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.678429 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39\": container with ID starting with bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39 not found: ID does not exist" containerID="bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.678475 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39"} err="failed to get container status \"bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39\": rpc error: code = NotFound desc = could not find container \"bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39\": container with ID starting with bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39 not found: ID does not exist" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.678507 4822 scope.go:117] "RemoveContainer" containerID="281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.679089 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8"} err="failed to get container status \"281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8\": rpc error: code = NotFound desc = could not find container \"281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8\": container with ID starting with 281216ef2015af97ca0fd7c02f194eeecac9d68fb83a45174ab7f30dba4eeae8 not found: ID does not exist" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.679112 4822 scope.go:117] "RemoveContainer" containerID="bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.680199 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39"} err="failed to get container status \"bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39\": rpc error: code = NotFound desc = could not find container \"bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39\": container with ID starting with bccec9a344bcd7e00454841b67436660c200c7c141dd5662a4add73f4d5f6d39 not found: ID does not exist" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.752942 4822 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9412c3-1984-48c1-9fa6-50cde5997fc7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.944409 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.969195 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.979491 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.979949 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e3bf9c-9f8b-4713-ba67-3e21635c4521" containerName="nova-manage" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.979966 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e3bf9c-9f8b-4713-ba67-3e21635c4521" containerName="nova-manage" Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.979999 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac64487-4fa7-42aa-a026-feedaaff7a94" containerName="init" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.980005 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac64487-4fa7-42aa-a026-feedaaff7a94" containerName="init" Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.980018 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerName="nova-api-api" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.980024 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerName="nova-api-api" Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.980032 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerName="nova-api-log" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.980037 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerName="nova-api-log" Nov 24 14:41:07 crc kubenswrapper[4822]: E1124 14:41:07.980050 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac64487-4fa7-42aa-a026-feedaaff7a94" containerName="dnsmasq-dns" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.980056 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac64487-4fa7-42aa-a026-feedaaff7a94" containerName="dnsmasq-dns" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.980266 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerName="nova-api-log" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.980280 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="75e3bf9c-9f8b-4713-ba67-3e21635c4521" containerName="nova-manage" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.980298 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac64487-4fa7-42aa-a026-feedaaff7a94" containerName="dnsmasq-dns" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.980317 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" containerName="nova-api-api" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.981456 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.987151 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.990907 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.990981 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 14:41:07 crc kubenswrapper[4822]: I1124 14:41:07.991069 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.161538 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-config-data\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.161838 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-internal-tls-certs\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.161867 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.162128 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-public-tls-certs\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.162292 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-logs\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.162588 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfs5v\" (UniqueName: \"kubernetes.io/projected/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-kube-api-access-cfs5v\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.264275 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-internal-tls-certs\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.264336 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.264499 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-public-tls-certs\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.264544 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-logs\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.264649 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfs5v\" (UniqueName: \"kubernetes.io/projected/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-kube-api-access-cfs5v\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.264703 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-config-data\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.265035 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-logs\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.269826 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.270481 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-internal-tls-certs\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.271957 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-public-tls-certs\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.272688 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-config-data\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.286874 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfs5v\" (UniqueName: \"kubernetes.io/projected/81b6cb5a-5c0f-49b8-ae7f-c3f81f553381-kube-api-access-cfs5v\") pod \"nova-api-0\" (UID: \"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381\") " pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.319831 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.592358 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.625044 4822 generic.go:334] "Generic (PLEG): container finished" podID="12c70ef7-2d2a-4251-9ad7-1d4bd631c399" containerID="ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe" exitCode=0 Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.625090 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"12c70ef7-2d2a-4251-9ad7-1d4bd631c399","Type":"ContainerDied","Data":"ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe"} Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.625125 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"12c70ef7-2d2a-4251-9ad7-1d4bd631c399","Type":"ContainerDied","Data":"a36952c61cb0cdd9a60a20ec775c391ba9b96ea9fe6ae4de0b3db72e63f6f2b3"} Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.625149 4822 scope.go:117] "RemoveContainer" containerID="ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.625169 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.658720 4822 scope.go:117] "RemoveContainer" containerID="ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe" Nov 24 14:41:08 crc kubenswrapper[4822]: E1124 14:41:08.659949 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe\": container with ID starting with ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe not found: ID does not exist" containerID="ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.660015 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe"} err="failed to get container status \"ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe\": rpc error: code = NotFound desc = could not find container \"ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe\": container with ID starting with ebaf9e8ba0d4c076e2f7e658dbdcf4f79d7ff4061587efa31fff1b172bd214fe not found: ID does not exist" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.777103 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-combined-ca-bundle\") pod \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.777191 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chtqb\" (UniqueName: \"kubernetes.io/projected/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-kube-api-access-chtqb\") pod \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.777330 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-config-data\") pod \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\" (UID: \"12c70ef7-2d2a-4251-9ad7-1d4bd631c399\") " Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.804327 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-kube-api-access-chtqb" (OuterVolumeSpecName: "kube-api-access-chtqb") pod "12c70ef7-2d2a-4251-9ad7-1d4bd631c399" (UID: "12c70ef7-2d2a-4251-9ad7-1d4bd631c399"). InnerVolumeSpecName "kube-api-access-chtqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.814314 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12c70ef7-2d2a-4251-9ad7-1d4bd631c399" (UID: "12c70ef7-2d2a-4251-9ad7-1d4bd631c399"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.835037 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-config-data" (OuterVolumeSpecName: "config-data") pod "12c70ef7-2d2a-4251-9ad7-1d4bd631c399" (UID: "12c70ef7-2d2a-4251-9ad7-1d4bd631c399"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.842690 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.880693 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.880728 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chtqb\" (UniqueName: \"kubernetes.io/projected/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-kube-api-access-chtqb\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:08 crc kubenswrapper[4822]: I1124 14:41:08.880743 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c70ef7-2d2a-4251-9ad7-1d4bd631c399-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.045368 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.059970 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.073355 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:41:09 crc kubenswrapper[4822]: E1124 14:41:09.074011 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c70ef7-2d2a-4251-9ad7-1d4bd631c399" containerName="nova-scheduler-scheduler" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.074045 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c70ef7-2d2a-4251-9ad7-1d4bd631c399" containerName="nova-scheduler-scheduler" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.074435 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c70ef7-2d2a-4251-9ad7-1d4bd631c399" containerName="nova-scheduler-scheduler" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.075843 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.078556 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.083753 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.192499 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa9ec181-a56f-4d25-925e-c5b9cf962cdf-config-data\") pod \"nova-scheduler-0\" (UID: \"fa9ec181-a56f-4d25-925e-c5b9cf962cdf\") " pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.192620 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa9ec181-a56f-4d25-925e-c5b9cf962cdf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fa9ec181-a56f-4d25-925e-c5b9cf962cdf\") " pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.192912 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvld7\" (UniqueName: \"kubernetes.io/projected/fa9ec181-a56f-4d25-925e-c5b9cf962cdf-kube-api-access-bvld7\") pod \"nova-scheduler-0\" (UID: \"fa9ec181-a56f-4d25-925e-c5b9cf962cdf\") " pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.295116 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa9ec181-a56f-4d25-925e-c5b9cf962cdf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fa9ec181-a56f-4d25-925e-c5b9cf962cdf\") " pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.295481 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvld7\" (UniqueName: \"kubernetes.io/projected/fa9ec181-a56f-4d25-925e-c5b9cf962cdf-kube-api-access-bvld7\") pod \"nova-scheduler-0\" (UID: \"fa9ec181-a56f-4d25-925e-c5b9cf962cdf\") " pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.295600 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa9ec181-a56f-4d25-925e-c5b9cf962cdf-config-data\") pod \"nova-scheduler-0\" (UID: \"fa9ec181-a56f-4d25-925e-c5b9cf962cdf\") " pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.300605 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa9ec181-a56f-4d25-925e-c5b9cf962cdf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fa9ec181-a56f-4d25-925e-c5b9cf962cdf\") " pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.303285 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa9ec181-a56f-4d25-925e-c5b9cf962cdf-config-data\") pod \"nova-scheduler-0\" (UID: \"fa9ec181-a56f-4d25-925e-c5b9cf962cdf\") " pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.314160 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvld7\" (UniqueName: \"kubernetes.io/projected/fa9ec181-a56f-4d25-925e-c5b9cf962cdf-kube-api-access-bvld7\") pod \"nova-scheduler-0\" (UID: \"fa9ec181-a56f-4d25-925e-c5b9cf962cdf\") " pod="openstack/nova-scheduler-0" Nov 24 14:41:09 crc kubenswrapper[4822]: I1124 14:41:09.394507 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:09.637666 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381","Type":"ContainerStarted","Data":"028d831f1d7b18343a3f8abe68537f9f781a945bfd36b4e4443d0991851472f3"} Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:09.638078 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381","Type":"ContainerStarted","Data":"52827448f4a930acbe2a8eaa1eea80b6cd9f44f46e94281ff90a11b155498e50"} Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:09.638088 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81b6cb5a-5c0f-49b8-ae7f-c3f81f553381","Type":"ContainerStarted","Data":"18b93406c446bbcab9626541ce96effa579760ee1bf44e00fe42fb3f97df8681"} Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:09.661907 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.661846273 podStartE2EDuration="2.661846273s" podCreationTimestamp="2025-11-24 14:41:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:09.655957436 +0000 UTC m=+1306.772597913" watchObservedRunningTime="2025-11-24 14:41:09.661846273 +0000 UTC m=+1306.778486770" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:09.722944 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c70ef7-2d2a-4251-9ad7-1d4bd631c399" path="/var/lib/kubelet/pods/12c70ef7-2d2a-4251-9ad7-1d4bd631c399/volumes" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:09.723694 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd9412c3-1984-48c1-9fa6-50cde5997fc7" path="/var/lib/kubelet/pods/dd9412c3-1984-48c1-9fa6-50cde5997fc7/volumes" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:09.883057 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:41:10 crc kubenswrapper[4822]: W1124 14:41:09.884906 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa9ec181_a56f_4d25_925e_c5b9cf962cdf.slice/crio-d013694a7aecbaeb9b55d8abcc859e21721f371001467dc15d833b4eae80ea2a WatchSource:0}: Error finding container d013694a7aecbaeb9b55d8abcc859e21721f371001467dc15d833b4eae80ea2a: Status 404 returned error can't find the container with id d013694a7aecbaeb9b55d8abcc859e21721f371001467dc15d833b4eae80ea2a Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:09.989896 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.216:8775/\": read tcp 10.217.0.2:57828->10.217.0.216:8775: read: connection reset by peer" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:09.989972 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.216:8775/\": read tcp 10.217.0.2:57834->10.217.0.216:8775: read: connection reset by peer" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.655527 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fa9ec181-a56f-4d25-925e-c5b9cf962cdf","Type":"ContainerStarted","Data":"f233437700733b973c585001e19739f26eb0ec64a53dd5806dd0fb48dcbe6bb7"} Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.655582 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fa9ec181-a56f-4d25-925e-c5b9cf962cdf","Type":"ContainerStarted","Data":"d013694a7aecbaeb9b55d8abcc859e21721f371001467dc15d833b4eae80ea2a"} Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.661092 4822 generic.go:334] "Generic (PLEG): container finished" podID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerID="e6533dec0b4a77a1cc2e167c2758c732b27d3561356b0dd94e42324fb97a8c32" exitCode=0 Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.661161 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0962a558-6bb1-44fe-953a-1e4f717b175e","Type":"ContainerDied","Data":"e6533dec0b4a77a1cc2e167c2758c732b27d3561356b0dd94e42324fb97a8c32"} Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.661375 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0962a558-6bb1-44fe-953a-1e4f717b175e","Type":"ContainerDied","Data":"8d7df195c3a162c730126881ed828cb4b803dc665532e26e7ff25a241ae2d5f3"} Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.661464 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d7df195c3a162c730126881ed828cb4b803dc665532e26e7ff25a241ae2d5f3" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.677940 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.67791337 podStartE2EDuration="1.67791337s" podCreationTimestamp="2025-11-24 14:41:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:10.674724748 +0000 UTC m=+1307.791365235" watchObservedRunningTime="2025-11-24 14:41:10.67791337 +0000 UTC m=+1307.794553867" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.706710 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.834008 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-nova-metadata-tls-certs\") pod \"0962a558-6bb1-44fe-953a-1e4f717b175e\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.834083 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-combined-ca-bundle\") pod \"0962a558-6bb1-44fe-953a-1e4f717b175e\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.834139 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2h52\" (UniqueName: \"kubernetes.io/projected/0962a558-6bb1-44fe-953a-1e4f717b175e-kube-api-access-f2h52\") pod \"0962a558-6bb1-44fe-953a-1e4f717b175e\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.834184 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-config-data\") pod \"0962a558-6bb1-44fe-953a-1e4f717b175e\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.834347 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0962a558-6bb1-44fe-953a-1e4f717b175e-logs\") pod \"0962a558-6bb1-44fe-953a-1e4f717b175e\" (UID: \"0962a558-6bb1-44fe-953a-1e4f717b175e\") " Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.835103 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0962a558-6bb1-44fe-953a-1e4f717b175e-logs" (OuterVolumeSpecName: "logs") pod "0962a558-6bb1-44fe-953a-1e4f717b175e" (UID: "0962a558-6bb1-44fe-953a-1e4f717b175e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.840650 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0962a558-6bb1-44fe-953a-1e4f717b175e-kube-api-access-f2h52" (OuterVolumeSpecName: "kube-api-access-f2h52") pod "0962a558-6bb1-44fe-953a-1e4f717b175e" (UID: "0962a558-6bb1-44fe-953a-1e4f717b175e"). InnerVolumeSpecName "kube-api-access-f2h52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.880300 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0962a558-6bb1-44fe-953a-1e4f717b175e" (UID: "0962a558-6bb1-44fe-953a-1e4f717b175e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.891574 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-config-data" (OuterVolumeSpecName: "config-data") pod "0962a558-6bb1-44fe-953a-1e4f717b175e" (UID: "0962a558-6bb1-44fe-953a-1e4f717b175e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.910847 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0962a558-6bb1-44fe-953a-1e4f717b175e" (UID: "0962a558-6bb1-44fe-953a-1e4f717b175e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.936951 4822 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.937190 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.937220 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2h52\" (UniqueName: \"kubernetes.io/projected/0962a558-6bb1-44fe-953a-1e4f717b175e-kube-api-access-f2h52\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.937232 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0962a558-6bb1-44fe-953a-1e4f717b175e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:10 crc kubenswrapper[4822]: I1124 14:41:10.937243 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0962a558-6bb1-44fe-953a-1e4f717b175e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.578335 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.578410 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.677035 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.739228 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.761446 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.780316 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:41:11 crc kubenswrapper[4822]: E1124 14:41:11.781397 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-log" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.781448 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-log" Nov 24 14:41:11 crc kubenswrapper[4822]: E1124 14:41:11.781511 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-metadata" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.781535 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-metadata" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.782055 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-log" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.782378 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" containerName="nova-metadata-metadata" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.785478 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.795785 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.796096 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.806709 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.962565 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d9bde3f-20ae-4497-8ce6-af4047c33364-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.962680 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9bde3f-20ae-4497-8ce6-af4047c33364-config-data\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.962720 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb4b7\" (UniqueName: \"kubernetes.io/projected/7d9bde3f-20ae-4497-8ce6-af4047c33364-kube-api-access-lb4b7\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.962740 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d9bde3f-20ae-4497-8ce6-af4047c33364-logs\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:11 crc kubenswrapper[4822]: I1124 14:41:11.962764 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9bde3f-20ae-4497-8ce6-af4047c33364-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.064288 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb4b7\" (UniqueName: \"kubernetes.io/projected/7d9bde3f-20ae-4497-8ce6-af4047c33364-kube-api-access-lb4b7\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.064345 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d9bde3f-20ae-4497-8ce6-af4047c33364-logs\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.064380 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9bde3f-20ae-4497-8ce6-af4047c33364-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.064527 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d9bde3f-20ae-4497-8ce6-af4047c33364-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.064633 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9bde3f-20ae-4497-8ce6-af4047c33364-config-data\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.065036 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d9bde3f-20ae-4497-8ce6-af4047c33364-logs\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.069033 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9bde3f-20ae-4497-8ce6-af4047c33364-config-data\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.070162 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9bde3f-20ae-4497-8ce6-af4047c33364-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.071558 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d9bde3f-20ae-4497-8ce6-af4047c33364-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.097344 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb4b7\" (UniqueName: \"kubernetes.io/projected/7d9bde3f-20ae-4497-8ce6-af4047c33364-kube-api-access-lb4b7\") pod \"nova-metadata-0\" (UID: \"7d9bde3f-20ae-4497-8ce6-af4047c33364\") " pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.123007 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:41:12 crc kubenswrapper[4822]: W1124 14:41:12.625468 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d9bde3f_20ae_4497_8ce6_af4047c33364.slice/crio-469b4c3bfe7aab63c6311923ca722e7a200bdb7bab2580d7d27ad717c2e77d36 WatchSource:0}: Error finding container 469b4c3bfe7aab63c6311923ca722e7a200bdb7bab2580d7d27ad717c2e77d36: Status 404 returned error can't find the container with id 469b4c3bfe7aab63c6311923ca722e7a200bdb7bab2580d7d27ad717c2e77d36 Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.628439 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:41:12 crc kubenswrapper[4822]: I1124 14:41:12.695008 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7d9bde3f-20ae-4497-8ce6-af4047c33364","Type":"ContainerStarted","Data":"469b4c3bfe7aab63c6311923ca722e7a200bdb7bab2580d7d27ad717c2e77d36"} Nov 24 14:41:13 crc kubenswrapper[4822]: I1124 14:41:13.723002 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0962a558-6bb1-44fe-953a-1e4f717b175e" path="/var/lib/kubelet/pods/0962a558-6bb1-44fe-953a-1e4f717b175e/volumes" Nov 24 14:41:13 crc kubenswrapper[4822]: I1124 14:41:13.724400 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7d9bde3f-20ae-4497-8ce6-af4047c33364","Type":"ContainerStarted","Data":"6f1d6fd9952958472d14a5d1ad53912faabc3ad751d68058aa22e84d261e6536"} Nov 24 14:41:13 crc kubenswrapper[4822]: I1124 14:41:13.724510 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7d9bde3f-20ae-4497-8ce6-af4047c33364","Type":"ContainerStarted","Data":"f08b09af6692a0d84e5bb6277d78d3883b6d84e3f8f5e03ad32ee63a649680ad"} Nov 24 14:41:13 crc kubenswrapper[4822]: I1124 14:41:13.772831 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.772809604 podStartE2EDuration="2.772809604s" podCreationTimestamp="2025-11-24 14:41:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:13.758764857 +0000 UTC m=+1310.875405344" watchObservedRunningTime="2025-11-24 14:41:13.772809604 +0000 UTC m=+1310.889450081" Nov 24 14:41:14 crc kubenswrapper[4822]: I1124 14:41:14.395586 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:41:17 crc kubenswrapper[4822]: I1124 14:41:17.124055 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:41:17 crc kubenswrapper[4822]: I1124 14:41:17.124481 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:41:18 crc kubenswrapper[4822]: I1124 14:41:18.321097 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:41:18 crc kubenswrapper[4822]: I1124 14:41:18.321582 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:41:19 crc kubenswrapper[4822]: I1124 14:41:19.341457 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81b6cb5a-5c0f-49b8-ae7f-c3f81f553381" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.226:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:41:19 crc kubenswrapper[4822]: I1124 14:41:19.341540 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81b6cb5a-5c0f-49b8-ae7f-c3f81f553381" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.226:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:41:19 crc kubenswrapper[4822]: I1124 14:41:19.394724 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 14:41:19 crc kubenswrapper[4822]: I1124 14:41:19.443468 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 14:41:19 crc kubenswrapper[4822]: I1124 14:41:19.876453 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 14:41:22 crc kubenswrapper[4822]: I1124 14:41:22.123709 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:41:22 crc kubenswrapper[4822]: I1124 14:41:22.124924 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:41:23 crc kubenswrapper[4822]: I1124 14:41:23.141481 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7d9bde3f-20ae-4497-8ce6-af4047c33364" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.228:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:41:23 crc kubenswrapper[4822]: I1124 14:41:23.141478 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7d9bde3f-20ae-4497-8ce6-af4047c33364" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.228:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:41:27 crc kubenswrapper[4822]: I1124 14:41:27.937256 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 14:41:28 crc kubenswrapper[4822]: I1124 14:41:28.330564 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:41:28 crc kubenswrapper[4822]: I1124 14:41:28.330992 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:41:28 crc kubenswrapper[4822]: I1124 14:41:28.331448 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:41:28 crc kubenswrapper[4822]: I1124 14:41:28.331520 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:41:28 crc kubenswrapper[4822]: I1124 14:41:28.338930 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:41:28 crc kubenswrapper[4822]: I1124 14:41:28.343155 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:41:32 crc kubenswrapper[4822]: I1124 14:41:32.130244 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:41:32 crc kubenswrapper[4822]: I1124 14:41:32.135969 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:41:32 crc kubenswrapper[4822]: I1124 14:41:32.142299 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:41:32 crc kubenswrapper[4822]: I1124 14:41:32.994635 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:41:41 crc kubenswrapper[4822]: I1124 14:41:41.578634 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:41:41 crc kubenswrapper[4822]: I1124 14:41:41.579109 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:42:11 crc kubenswrapper[4822]: I1124 14:42:11.577876 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:42:11 crc kubenswrapper[4822]: I1124 14:42:11.579439 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:42:11 crc kubenswrapper[4822]: I1124 14:42:11.579541 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:42:11 crc kubenswrapper[4822]: I1124 14:42:11.580508 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9b7bf9be06a9bc314261db2c2d98effa00916294896221883633ba00a47f5d1"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:42:11 crc kubenswrapper[4822]: I1124 14:42:11.580572 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://a9b7bf9be06a9bc314261db2c2d98effa00916294896221883633ba00a47f5d1" gracePeriod=600 Nov 24 14:42:12 crc kubenswrapper[4822]: I1124 14:42:12.508737 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="a9b7bf9be06a9bc314261db2c2d98effa00916294896221883633ba00a47f5d1" exitCode=0 Nov 24 14:42:12 crc kubenswrapper[4822]: I1124 14:42:12.508803 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"a9b7bf9be06a9bc314261db2c2d98effa00916294896221883633ba00a47f5d1"} Nov 24 14:42:12 crc kubenswrapper[4822]: I1124 14:42:12.509522 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678"} Nov 24 14:42:12 crc kubenswrapper[4822]: I1124 14:42:12.509608 4822 scope.go:117] "RemoveContainer" containerID="eb98f980c90c3e30ffa6cd4e582af8567a0ef6fab6806bdeaf156e64abf982bb" Nov 24 14:42:25 crc kubenswrapper[4822]: I1124 14:42:25.716401 4822 scope.go:117] "RemoveContainer" containerID="21835380714c9cbcba6c8677e016f4c511748a6456dd539a1b8d0b8e22f07f16" Nov 24 14:42:25 crc kubenswrapper[4822]: I1124 14:42:25.746327 4822 scope.go:117] "RemoveContainer" containerID="8bd88d4804ede327a228082bc760bbc64ceb54ce2bcdee57361364ea7ef45712" Nov 24 14:42:25 crc kubenswrapper[4822]: I1124 14:42:25.774504 4822 scope.go:117] "RemoveContainer" containerID="7112ed58314143939d290e79da2a1799438f73664f74b747fe08285adc029d52" Nov 24 14:42:47 crc kubenswrapper[4822]: E1124 14:42:47.173649 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="4acc7e6a-472b-468a-b709-183f8b3c2b5b" Nov 24 14:42:47 crc kubenswrapper[4822]: I1124 14:42:47.985157 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:42:49 crc kubenswrapper[4822]: I1124 14:42:49.146709 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:42:49 crc kubenswrapper[4822]: E1124 14:42:49.146959 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:42:49 crc kubenswrapper[4822]: E1124 14:42:49.146989 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:42:49 crc kubenswrapper[4822]: E1124 14:42:49.147069 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:44:51.147044694 +0000 UTC m=+1528.263685211 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:42:49 crc kubenswrapper[4822]: I1124 14:42:49.879962 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jbqwc"] Nov 24 14:42:49 crc kubenswrapper[4822]: I1124 14:42:49.886286 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:49 crc kubenswrapper[4822]: I1124 14:42:49.900620 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jbqwc"] Nov 24 14:42:49 crc kubenswrapper[4822]: I1124 14:42:49.971128 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nshh4\" (UniqueName: \"kubernetes.io/projected/114f5954-3731-45c4-8e98-afaf4170f630-kube-api-access-nshh4\") pod \"certified-operators-jbqwc\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:49 crc kubenswrapper[4822]: I1124 14:42:49.971181 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-catalog-content\") pod \"certified-operators-jbqwc\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:49 crc kubenswrapper[4822]: I1124 14:42:49.971288 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-utilities\") pod \"certified-operators-jbqwc\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:50 crc kubenswrapper[4822]: I1124 14:42:50.073542 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-utilities\") pod \"certified-operators-jbqwc\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:50 crc kubenswrapper[4822]: I1124 14:42:50.073792 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nshh4\" (UniqueName: \"kubernetes.io/projected/114f5954-3731-45c4-8e98-afaf4170f630-kube-api-access-nshh4\") pod \"certified-operators-jbqwc\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:50 crc kubenswrapper[4822]: I1124 14:42:50.073818 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-catalog-content\") pod \"certified-operators-jbqwc\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:50 crc kubenswrapper[4822]: I1124 14:42:50.074520 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-catalog-content\") pod \"certified-operators-jbqwc\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:50 crc kubenswrapper[4822]: I1124 14:42:50.074913 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-utilities\") pod \"certified-operators-jbqwc\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:50 crc kubenswrapper[4822]: I1124 14:42:50.104693 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nshh4\" (UniqueName: \"kubernetes.io/projected/114f5954-3731-45c4-8e98-afaf4170f630-kube-api-access-nshh4\") pod \"certified-operators-jbqwc\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:50 crc kubenswrapper[4822]: I1124 14:42:50.222680 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:42:50 crc kubenswrapper[4822]: I1124 14:42:50.780359 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jbqwc"] Nov 24 14:42:51 crc kubenswrapper[4822]: I1124 14:42:51.020958 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbqwc" event={"ID":"114f5954-3731-45c4-8e98-afaf4170f630","Type":"ContainerStarted","Data":"26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54"} Nov 24 14:42:51 crc kubenswrapper[4822]: I1124 14:42:51.021033 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbqwc" event={"ID":"114f5954-3731-45c4-8e98-afaf4170f630","Type":"ContainerStarted","Data":"26850cb536aa636cc60b46abaa20f0f51d2edad45fa93020b9e5e349e7c42b5d"} Nov 24 14:42:52 crc kubenswrapper[4822]: I1124 14:42:52.035844 4822 generic.go:334] "Generic (PLEG): container finished" podID="114f5954-3731-45c4-8e98-afaf4170f630" containerID="26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54" exitCode=0 Nov 24 14:42:52 crc kubenswrapper[4822]: I1124 14:42:52.035951 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbqwc" event={"ID":"114f5954-3731-45c4-8e98-afaf4170f630","Type":"ContainerDied","Data":"26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54"} Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.457063 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-49c8x"] Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.462375 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.469593 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-49c8x"] Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.567076 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-utilities\") pod \"community-operators-49c8x\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.567302 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx7b2\" (UniqueName: \"kubernetes.io/projected/8b59173b-bf14-44cb-8bf3-8772691fb87a-kube-api-access-kx7b2\") pod \"community-operators-49c8x\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.567448 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-catalog-content\") pod \"community-operators-49c8x\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.670088 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-utilities\") pod \"community-operators-49c8x\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.670171 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx7b2\" (UniqueName: \"kubernetes.io/projected/8b59173b-bf14-44cb-8bf3-8772691fb87a-kube-api-access-kx7b2\") pod \"community-operators-49c8x\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.670257 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-catalog-content\") pod \"community-operators-49c8x\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.670835 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-catalog-content\") pod \"community-operators-49c8x\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.670844 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-utilities\") pod \"community-operators-49c8x\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.688730 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx7b2\" (UniqueName: \"kubernetes.io/projected/8b59173b-bf14-44cb-8bf3-8772691fb87a-kube-api-access-kx7b2\") pod \"community-operators-49c8x\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:53 crc kubenswrapper[4822]: I1124 14:42:53.790818 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:42:54 crc kubenswrapper[4822]: I1124 14:42:54.311022 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-49c8x"] Nov 24 14:42:54 crc kubenswrapper[4822]: W1124 14:42:54.312376 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b59173b_bf14_44cb_8bf3_8772691fb87a.slice/crio-8cfb17e9bc9f227d970648950f476ecdd2d8a05d696c9e3fd8dbb5e3539b1786 WatchSource:0}: Error finding container 8cfb17e9bc9f227d970648950f476ecdd2d8a05d696c9e3fd8dbb5e3539b1786: Status 404 returned error can't find the container with id 8cfb17e9bc9f227d970648950f476ecdd2d8a05d696c9e3fd8dbb5e3539b1786 Nov 24 14:42:55 crc kubenswrapper[4822]: I1124 14:42:55.092011 4822 generic.go:334] "Generic (PLEG): container finished" podID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerID="bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c" exitCode=0 Nov 24 14:42:55 crc kubenswrapper[4822]: I1124 14:42:55.092111 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49c8x" event={"ID":"8b59173b-bf14-44cb-8bf3-8772691fb87a","Type":"ContainerDied","Data":"bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c"} Nov 24 14:42:55 crc kubenswrapper[4822]: I1124 14:42:55.092404 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49c8x" event={"ID":"8b59173b-bf14-44cb-8bf3-8772691fb87a","Type":"ContainerStarted","Data":"8cfb17e9bc9f227d970648950f476ecdd2d8a05d696c9e3fd8dbb5e3539b1786"} Nov 24 14:42:58 crc kubenswrapper[4822]: I1124 14:42:58.132107 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49c8x" event={"ID":"8b59173b-bf14-44cb-8bf3-8772691fb87a","Type":"ContainerStarted","Data":"654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e"} Nov 24 14:42:58 crc kubenswrapper[4822]: I1124 14:42:58.135538 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbqwc" event={"ID":"114f5954-3731-45c4-8e98-afaf4170f630","Type":"ContainerStarted","Data":"499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208"} Nov 24 14:42:59 crc kubenswrapper[4822]: I1124 14:42:59.157696 4822 generic.go:334] "Generic (PLEG): container finished" podID="114f5954-3731-45c4-8e98-afaf4170f630" containerID="499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208" exitCode=0 Nov 24 14:42:59 crc kubenswrapper[4822]: I1124 14:42:59.157887 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbqwc" event={"ID":"114f5954-3731-45c4-8e98-afaf4170f630","Type":"ContainerDied","Data":"499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208"} Nov 24 14:42:59 crc kubenswrapper[4822]: I1124 14:42:59.163195 4822 generic.go:334] "Generic (PLEG): container finished" podID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerID="654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e" exitCode=0 Nov 24 14:42:59 crc kubenswrapper[4822]: I1124 14:42:59.163266 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49c8x" event={"ID":"8b59173b-bf14-44cb-8bf3-8772691fb87a","Type":"ContainerDied","Data":"654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e"} Nov 24 14:43:00 crc kubenswrapper[4822]: I1124 14:43:00.174646 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbqwc" event={"ID":"114f5954-3731-45c4-8e98-afaf4170f630","Type":"ContainerStarted","Data":"ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47"} Nov 24 14:43:00 crc kubenswrapper[4822]: I1124 14:43:00.176927 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49c8x" event={"ID":"8b59173b-bf14-44cb-8bf3-8772691fb87a","Type":"ContainerStarted","Data":"14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002"} Nov 24 14:43:00 crc kubenswrapper[4822]: I1124 14:43:00.200111 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jbqwc" podStartSLOduration=3.55284475 podStartE2EDuration="11.200088169s" podCreationTimestamp="2025-11-24 14:42:49 +0000 UTC" firstStartedPulling="2025-11-24 14:42:52.037981438 +0000 UTC m=+1409.154621915" lastFinishedPulling="2025-11-24 14:42:59.685224857 +0000 UTC m=+1416.801865334" observedRunningTime="2025-11-24 14:43:00.190164468 +0000 UTC m=+1417.306804945" watchObservedRunningTime="2025-11-24 14:43:00.200088169 +0000 UTC m=+1417.316728646" Nov 24 14:43:00 crc kubenswrapper[4822]: I1124 14:43:00.215086 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-49c8x" podStartSLOduration=2.707462628 podStartE2EDuration="7.215062769s" podCreationTimestamp="2025-11-24 14:42:53 +0000 UTC" firstStartedPulling="2025-11-24 14:42:55.095068304 +0000 UTC m=+1412.211708791" lastFinishedPulling="2025-11-24 14:42:59.602668435 +0000 UTC m=+1416.719308932" observedRunningTime="2025-11-24 14:43:00.211867639 +0000 UTC m=+1417.328508126" watchObservedRunningTime="2025-11-24 14:43:00.215062769 +0000 UTC m=+1417.331703246" Nov 24 14:43:00 crc kubenswrapper[4822]: I1124 14:43:00.224138 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:43:00 crc kubenswrapper[4822]: I1124 14:43:00.224195 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:43:01 crc kubenswrapper[4822]: I1124 14:43:01.288223 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jbqwc" podUID="114f5954-3731-45c4-8e98-afaf4170f630" containerName="registry-server" probeResult="failure" output=< Nov 24 14:43:01 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:43:01 crc kubenswrapper[4822]: > Nov 24 14:43:03 crc kubenswrapper[4822]: I1124 14:43:03.791828 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:43:03 crc kubenswrapper[4822]: I1124 14:43:03.792201 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:43:03 crc kubenswrapper[4822]: I1124 14:43:03.866174 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:43:04 crc kubenswrapper[4822]: I1124 14:43:04.317369 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:43:04 crc kubenswrapper[4822]: I1124 14:43:04.447283 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-49c8x"] Nov 24 14:43:05 crc kubenswrapper[4822]: E1124 14:43:05.497403 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" podUID="753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf" Nov 24 14:43:06 crc kubenswrapper[4822]: I1124 14:43:06.263192 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:43:06 crc kubenswrapper[4822]: I1124 14:43:06.263681 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-49c8x" podUID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerName="registry-server" containerID="cri-o://14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002" gracePeriod=2 Nov 24 14:43:06 crc kubenswrapper[4822]: I1124 14:43:06.878974 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.014916 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-utilities\") pod \"8b59173b-bf14-44cb-8bf3-8772691fb87a\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.015383 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-catalog-content\") pod \"8b59173b-bf14-44cb-8bf3-8772691fb87a\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.015638 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx7b2\" (UniqueName: \"kubernetes.io/projected/8b59173b-bf14-44cb-8bf3-8772691fb87a-kube-api-access-kx7b2\") pod \"8b59173b-bf14-44cb-8bf3-8772691fb87a\" (UID: \"8b59173b-bf14-44cb-8bf3-8772691fb87a\") " Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.016114 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-utilities" (OuterVolumeSpecName: "utilities") pod "8b59173b-bf14-44cb-8bf3-8772691fb87a" (UID: "8b59173b-bf14-44cb-8bf3-8772691fb87a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.016513 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.026844 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b59173b-bf14-44cb-8bf3-8772691fb87a-kube-api-access-kx7b2" (OuterVolumeSpecName: "kube-api-access-kx7b2") pod "8b59173b-bf14-44cb-8bf3-8772691fb87a" (UID: "8b59173b-bf14-44cb-8bf3-8772691fb87a"). InnerVolumeSpecName "kube-api-access-kx7b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.082097 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b59173b-bf14-44cb-8bf3-8772691fb87a" (UID: "8b59173b-bf14-44cb-8bf3-8772691fb87a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.118039 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b59173b-bf14-44cb-8bf3-8772691fb87a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.118069 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx7b2\" (UniqueName: \"kubernetes.io/projected/8b59173b-bf14-44cb-8bf3-8772691fb87a-kube-api-access-kx7b2\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.273080 4822 generic.go:334] "Generic (PLEG): container finished" podID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerID="14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002" exitCode=0 Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.273130 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49c8x" event={"ID":"8b59173b-bf14-44cb-8bf3-8772691fb87a","Type":"ContainerDied","Data":"14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002"} Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.273177 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49c8x" event={"ID":"8b59173b-bf14-44cb-8bf3-8772691fb87a","Type":"ContainerDied","Data":"8cfb17e9bc9f227d970648950f476ecdd2d8a05d696c9e3fd8dbb5e3539b1786"} Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.273204 4822 scope.go:117] "RemoveContainer" containerID="14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.273222 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49c8x" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.294719 4822 scope.go:117] "RemoveContainer" containerID="654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.306617 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-49c8x"] Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.315771 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-49c8x"] Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.336761 4822 scope.go:117] "RemoveContainer" containerID="bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.418738 4822 scope.go:117] "RemoveContainer" containerID="14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002" Nov 24 14:43:07 crc kubenswrapper[4822]: E1124 14:43:07.419095 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002\": container with ID starting with 14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002 not found: ID does not exist" containerID="14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.419135 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002"} err="failed to get container status \"14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002\": rpc error: code = NotFound desc = could not find container \"14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002\": container with ID starting with 14037ff99fa9a7ffb310dcb5073e86f50dcedb1542cff42d33b43eb0f8f16002 not found: ID does not exist" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.419158 4822 scope.go:117] "RemoveContainer" containerID="654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e" Nov 24 14:43:07 crc kubenswrapper[4822]: E1124 14:43:07.419504 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e\": container with ID starting with 654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e not found: ID does not exist" containerID="654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.419527 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e"} err="failed to get container status \"654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e\": rpc error: code = NotFound desc = could not find container \"654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e\": container with ID starting with 654f658fc07419a0741dc95f595cb7cb1b19dd63230ec1eb54501422d8168d9e not found: ID does not exist" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.419543 4822 scope.go:117] "RemoveContainer" containerID="bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c" Nov 24 14:43:07 crc kubenswrapper[4822]: E1124 14:43:07.419763 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c\": container with ID starting with bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c not found: ID does not exist" containerID="bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.419789 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c"} err="failed to get container status \"bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c\": rpc error: code = NotFound desc = could not find container \"bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c\": container with ID starting with bfeedda1fde1e8d3f82f408be7af5fc7f0bb9d2ab6234445e05f16c2fd8fdd8c not found: ID does not exist" Nov 24 14:43:07 crc kubenswrapper[4822]: I1124 14:43:07.719389 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b59173b-bf14-44cb-8bf3-8772691fb87a" path="/var/lib/kubelet/pods/8b59173b-bf14-44cb-8bf3-8772691fb87a/volumes" Nov 24 14:43:09 crc kubenswrapper[4822]: I1124 14:43:09.475511 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:43:09 crc kubenswrapper[4822]: E1124 14:43:09.475763 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:43:09 crc kubenswrapper[4822]: E1124 14:43:09.475796 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:43:09 crc kubenswrapper[4822]: E1124 14:43:09.475866 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:45:11.475847051 +0000 UTC m=+1548.592487528 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:43:10 crc kubenswrapper[4822]: I1124 14:43:10.304703 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:43:10 crc kubenswrapper[4822]: I1124 14:43:10.359775 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:43:10 crc kubenswrapper[4822]: I1124 14:43:10.861067 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jbqwc"] Nov 24 14:43:12 crc kubenswrapper[4822]: I1124 14:43:12.328460 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jbqwc" podUID="114f5954-3731-45c4-8e98-afaf4170f630" containerName="registry-server" containerID="cri-o://ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47" gracePeriod=2 Nov 24 14:43:12 crc kubenswrapper[4822]: I1124 14:43:12.935106 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.057355 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nshh4\" (UniqueName: \"kubernetes.io/projected/114f5954-3731-45c4-8e98-afaf4170f630-kube-api-access-nshh4\") pod \"114f5954-3731-45c4-8e98-afaf4170f630\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.057583 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-utilities\") pod \"114f5954-3731-45c4-8e98-afaf4170f630\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.057616 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-catalog-content\") pod \"114f5954-3731-45c4-8e98-afaf4170f630\" (UID: \"114f5954-3731-45c4-8e98-afaf4170f630\") " Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.058513 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-utilities" (OuterVolumeSpecName: "utilities") pod "114f5954-3731-45c4-8e98-afaf4170f630" (UID: "114f5954-3731-45c4-8e98-afaf4170f630"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.063944 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/114f5954-3731-45c4-8e98-afaf4170f630-kube-api-access-nshh4" (OuterVolumeSpecName: "kube-api-access-nshh4") pod "114f5954-3731-45c4-8e98-afaf4170f630" (UID: "114f5954-3731-45c4-8e98-afaf4170f630"). InnerVolumeSpecName "kube-api-access-nshh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.124849 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "114f5954-3731-45c4-8e98-afaf4170f630" (UID: "114f5954-3731-45c4-8e98-afaf4170f630"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.160990 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.161022 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/114f5954-3731-45c4-8e98-afaf4170f630-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.161038 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nshh4\" (UniqueName: \"kubernetes.io/projected/114f5954-3731-45c4-8e98-afaf4170f630-kube-api-access-nshh4\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.358632 4822 generic.go:334] "Generic (PLEG): container finished" podID="114f5954-3731-45c4-8e98-afaf4170f630" containerID="ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47" exitCode=0 Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.358679 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbqwc" event={"ID":"114f5954-3731-45c4-8e98-afaf4170f630","Type":"ContainerDied","Data":"ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47"} Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.358708 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbqwc" event={"ID":"114f5954-3731-45c4-8e98-afaf4170f630","Type":"ContainerDied","Data":"26850cb536aa636cc60b46abaa20f0f51d2edad45fa93020b9e5e349e7c42b5d"} Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.358728 4822 scope.go:117] "RemoveContainer" containerID="ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.358725 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbqwc" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.413835 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jbqwc"] Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.420501 4822 scope.go:117] "RemoveContainer" containerID="499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.425062 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jbqwc"] Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.444282 4822 scope.go:117] "RemoveContainer" containerID="26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.489006 4822 scope.go:117] "RemoveContainer" containerID="ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47" Nov 24 14:43:13 crc kubenswrapper[4822]: E1124 14:43:13.489413 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47\": container with ID starting with ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47 not found: ID does not exist" containerID="ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.489446 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47"} err="failed to get container status \"ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47\": rpc error: code = NotFound desc = could not find container \"ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47\": container with ID starting with ac2537c7c9d3fecaf60c23821f976f35cf59f612ae0947bc8a713d680fc4eb47 not found: ID does not exist" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.489467 4822 scope.go:117] "RemoveContainer" containerID="499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208" Nov 24 14:43:13 crc kubenswrapper[4822]: E1124 14:43:13.489777 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208\": container with ID starting with 499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208 not found: ID does not exist" containerID="499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.489798 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208"} err="failed to get container status \"499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208\": rpc error: code = NotFound desc = could not find container \"499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208\": container with ID starting with 499b38f05f14032c968900c4203a02b4186844ec2d8caa5a361fb757f1120208 not found: ID does not exist" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.489814 4822 scope.go:117] "RemoveContainer" containerID="26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54" Nov 24 14:43:13 crc kubenswrapper[4822]: E1124 14:43:13.490034 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54\": container with ID starting with 26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54 not found: ID does not exist" containerID="26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.490053 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54"} err="failed to get container status \"26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54\": rpc error: code = NotFound desc = could not find container \"26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54\": container with ID starting with 26a355653aa1196bedf6a2a75fb3e845a7e25f209b336db33a6010dbbbefae54 not found: ID does not exist" Nov 24 14:43:13 crc kubenswrapper[4822]: I1124 14:43:13.725439 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="114f5954-3731-45c4-8e98-afaf4170f630" path="/var/lib/kubelet/pods/114f5954-3731-45c4-8e98-afaf4170f630/volumes" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.346686 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hhwnx"] Nov 24 14:43:17 crc kubenswrapper[4822]: E1124 14:43:17.347945 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="114f5954-3731-45c4-8e98-afaf4170f630" containerName="extract-content" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.347970 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="114f5954-3731-45c4-8e98-afaf4170f630" containerName="extract-content" Nov 24 14:43:17 crc kubenswrapper[4822]: E1124 14:43:17.348000 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerName="registry-server" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.348013 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerName="registry-server" Nov 24 14:43:17 crc kubenswrapper[4822]: E1124 14:43:17.348038 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerName="extract-utilities" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.348054 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerName="extract-utilities" Nov 24 14:43:17 crc kubenswrapper[4822]: E1124 14:43:17.348099 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="114f5954-3731-45c4-8e98-afaf4170f630" containerName="extract-utilities" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.348115 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="114f5954-3731-45c4-8e98-afaf4170f630" containerName="extract-utilities" Nov 24 14:43:17 crc kubenswrapper[4822]: E1124 14:43:17.348137 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="114f5954-3731-45c4-8e98-afaf4170f630" containerName="registry-server" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.348149 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="114f5954-3731-45c4-8e98-afaf4170f630" containerName="registry-server" Nov 24 14:43:17 crc kubenswrapper[4822]: E1124 14:43:17.348182 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerName="extract-content" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.348195 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerName="extract-content" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.348627 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="114f5954-3731-45c4-8e98-afaf4170f630" containerName="registry-server" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.348660 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b59173b-bf14-44cb-8bf3-8772691fb87a" containerName="registry-server" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.351395 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.362382 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhwnx"] Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.475475 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-catalog-content\") pod \"redhat-marketplace-hhwnx\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.475635 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-utilities\") pod \"redhat-marketplace-hhwnx\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.475744 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szhl2\" (UniqueName: \"kubernetes.io/projected/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-kube-api-access-szhl2\") pod \"redhat-marketplace-hhwnx\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.578807 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-catalog-content\") pod \"redhat-marketplace-hhwnx\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.578927 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-utilities\") pod \"redhat-marketplace-hhwnx\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.579033 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szhl2\" (UniqueName: \"kubernetes.io/projected/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-kube-api-access-szhl2\") pod \"redhat-marketplace-hhwnx\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.579367 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-catalog-content\") pod \"redhat-marketplace-hhwnx\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.579600 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-utilities\") pod \"redhat-marketplace-hhwnx\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.601891 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szhl2\" (UniqueName: \"kubernetes.io/projected/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-kube-api-access-szhl2\") pod \"redhat-marketplace-hhwnx\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:17 crc kubenswrapper[4822]: I1124 14:43:17.721332 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:18 crc kubenswrapper[4822]: I1124 14:43:18.338956 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhwnx"] Nov 24 14:43:18 crc kubenswrapper[4822]: I1124 14:43:18.442186 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhwnx" event={"ID":"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53","Type":"ContainerStarted","Data":"c668fa11188cfe7e75acf2eacb41c8bbe744c17d82e2d86b349b8a3b5f32e31b"} Nov 24 14:43:19 crc kubenswrapper[4822]: I1124 14:43:19.461788 4822 generic.go:334] "Generic (PLEG): container finished" podID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerID="fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a" exitCode=0 Nov 24 14:43:19 crc kubenswrapper[4822]: I1124 14:43:19.461874 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhwnx" event={"ID":"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53","Type":"ContainerDied","Data":"fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a"} Nov 24 14:43:20 crc kubenswrapper[4822]: I1124 14:43:20.476383 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhwnx" event={"ID":"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53","Type":"ContainerStarted","Data":"6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab"} Nov 24 14:43:21 crc kubenswrapper[4822]: I1124 14:43:21.490312 4822 generic.go:334] "Generic (PLEG): container finished" podID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerID="6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab" exitCode=0 Nov 24 14:43:21 crc kubenswrapper[4822]: I1124 14:43:21.490469 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhwnx" event={"ID":"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53","Type":"ContainerDied","Data":"6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab"} Nov 24 14:43:22 crc kubenswrapper[4822]: I1124 14:43:22.504561 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhwnx" event={"ID":"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53","Type":"ContainerStarted","Data":"a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f"} Nov 24 14:43:22 crc kubenswrapper[4822]: I1124 14:43:22.530604 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hhwnx" podStartSLOduration=3.120359809 podStartE2EDuration="5.530582126s" podCreationTimestamp="2025-11-24 14:43:17 +0000 UTC" firstStartedPulling="2025-11-24 14:43:19.466281914 +0000 UTC m=+1436.582922431" lastFinishedPulling="2025-11-24 14:43:21.876504271 +0000 UTC m=+1438.993144748" observedRunningTime="2025-11-24 14:43:22.528127679 +0000 UTC m=+1439.644768196" watchObservedRunningTime="2025-11-24 14:43:22.530582126 +0000 UTC m=+1439.647222613" Nov 24 14:43:25 crc kubenswrapper[4822]: I1124 14:43:25.940644 4822 scope.go:117] "RemoveContainer" containerID="f417c33072a5a4a65267ea26ba787c959ddfcb24d817afa0933cd5bf9cd1d067" Nov 24 14:43:27 crc kubenswrapper[4822]: I1124 14:43:27.727528 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:27 crc kubenswrapper[4822]: I1124 14:43:27.728024 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:27 crc kubenswrapper[4822]: I1124 14:43:27.814073 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:28 crc kubenswrapper[4822]: I1124 14:43:28.666048 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:28 crc kubenswrapper[4822]: I1124 14:43:28.730851 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhwnx"] Nov 24 14:43:30 crc kubenswrapper[4822]: I1124 14:43:30.606906 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hhwnx" podUID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerName="registry-server" containerID="cri-o://a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f" gracePeriod=2 Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.298642 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.404428 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-catalog-content\") pod \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.404705 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szhl2\" (UniqueName: \"kubernetes.io/projected/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-kube-api-access-szhl2\") pod \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.404853 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-utilities\") pod \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\" (UID: \"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53\") " Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.405697 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-utilities" (OuterVolumeSpecName: "utilities") pod "b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" (UID: "b59b2e25-2f68-460c-b4a2-f09e3ef0aa53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.406517 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.413422 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-kube-api-access-szhl2" (OuterVolumeSpecName: "kube-api-access-szhl2") pod "b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" (UID: "b59b2e25-2f68-460c-b4a2-f09e3ef0aa53"). InnerVolumeSpecName "kube-api-access-szhl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.427876 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" (UID: "b59b2e25-2f68-460c-b4a2-f09e3ef0aa53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.507834 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.507875 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szhl2\" (UniqueName: \"kubernetes.io/projected/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53-kube-api-access-szhl2\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.634817 4822 generic.go:334] "Generic (PLEG): container finished" podID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerID="a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f" exitCode=0 Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.634869 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhwnx" event={"ID":"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53","Type":"ContainerDied","Data":"a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f"} Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.634903 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhwnx" event={"ID":"b59b2e25-2f68-460c-b4a2-f09e3ef0aa53","Type":"ContainerDied","Data":"c668fa11188cfe7e75acf2eacb41c8bbe744c17d82e2d86b349b8a3b5f32e31b"} Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.634923 4822 scope.go:117] "RemoveContainer" containerID="a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.635100 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhwnx" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.665479 4822 scope.go:117] "RemoveContainer" containerID="6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.736373 4822 scope.go:117] "RemoveContainer" containerID="fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.738435 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhwnx"] Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.738464 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhwnx"] Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.789665 4822 scope.go:117] "RemoveContainer" containerID="a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f" Nov 24 14:43:31 crc kubenswrapper[4822]: E1124 14:43:31.793309 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f\": container with ID starting with a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f not found: ID does not exist" containerID="a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.793352 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f"} err="failed to get container status \"a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f\": rpc error: code = NotFound desc = could not find container \"a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f\": container with ID starting with a3644ea2e4217ed9c05404054917cff324ce1399f06408aed9ccef5b2adf928f not found: ID does not exist" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.793381 4822 scope.go:117] "RemoveContainer" containerID="6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab" Nov 24 14:43:31 crc kubenswrapper[4822]: E1124 14:43:31.797291 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab\": container with ID starting with 6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab not found: ID does not exist" containerID="6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.797320 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab"} err="failed to get container status \"6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab\": rpc error: code = NotFound desc = could not find container \"6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab\": container with ID starting with 6a7b4528ef4927883404e916e059b7fd4a4b8ef34b2e3f1d7d3ad04d8fac2fab not found: ID does not exist" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.797335 4822 scope.go:117] "RemoveContainer" containerID="fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a" Nov 24 14:43:31 crc kubenswrapper[4822]: E1124 14:43:31.801286 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a\": container with ID starting with fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a not found: ID does not exist" containerID="fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a" Nov 24 14:43:31 crc kubenswrapper[4822]: I1124 14:43:31.801311 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a"} err="failed to get container status \"fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a\": rpc error: code = NotFound desc = could not find container \"fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a\": container with ID starting with fbd312618f782fbc697acea013bdd43940efda35091f106ad791bb786557945a not found: ID does not exist" Nov 24 14:43:33 crc kubenswrapper[4822]: I1124 14:43:33.718963 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" path="/var/lib/kubelet/pods/b59b2e25-2f68-460c-b4a2-f09e3ef0aa53/volumes" Nov 24 14:44:11 crc kubenswrapper[4822]: I1124 14:44:11.578507 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:44:11 crc kubenswrapper[4822]: I1124 14:44:11.580533 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:44:26 crc kubenswrapper[4822]: I1124 14:44:26.082356 4822 scope.go:117] "RemoveContainer" containerID="478c0129945058472624762055c6cbbcff7171196e7b8e7b05755afd0d7ab513" Nov 24 14:44:41 crc kubenswrapper[4822]: I1124 14:44:41.578117 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:44:41 crc kubenswrapper[4822]: I1124 14:44:41.578749 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:44:50 crc kubenswrapper[4822]: E1124 14:44:50.988792 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="4acc7e6a-472b-468a-b709-183f8b3c2b5b" Nov 24 14:44:51 crc kubenswrapper[4822]: I1124 14:44:51.216310 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:44:51 crc kubenswrapper[4822]: E1124 14:44:51.216638 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:44:51 crc kubenswrapper[4822]: E1124 14:44:51.216708 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:44:51 crc kubenswrapper[4822]: E1124 14:44:51.216818 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:46:53.216782509 +0000 UTC m=+1650.333423026 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:44:51 crc kubenswrapper[4822]: I1124 14:44:51.670595 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.173349 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6"] Nov 24 14:45:00 crc kubenswrapper[4822]: E1124 14:45:00.174869 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerName="extract-utilities" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.174921 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerName="extract-utilities" Nov 24 14:45:00 crc kubenswrapper[4822]: E1124 14:45:00.174951 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerName="registry-server" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.174962 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerName="registry-server" Nov 24 14:45:00 crc kubenswrapper[4822]: E1124 14:45:00.175043 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerName="extract-content" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.175057 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerName="extract-content" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.175735 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="b59b2e25-2f68-460c-b4a2-f09e3ef0aa53" containerName="registry-server" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.177229 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.180720 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.181713 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.186869 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6"] Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.244327 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llm4j\" (UniqueName: \"kubernetes.io/projected/797e852a-cd41-4cd8-b2da-89c2407d14f2-kube-api-access-llm4j\") pod \"collect-profiles-29399925-x74s6\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.244369 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/797e852a-cd41-4cd8-b2da-89c2407d14f2-config-volume\") pod \"collect-profiles-29399925-x74s6\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.244387 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/797e852a-cd41-4cd8-b2da-89c2407d14f2-secret-volume\") pod \"collect-profiles-29399925-x74s6\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.346316 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llm4j\" (UniqueName: \"kubernetes.io/projected/797e852a-cd41-4cd8-b2da-89c2407d14f2-kube-api-access-llm4j\") pod \"collect-profiles-29399925-x74s6\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.346379 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/797e852a-cd41-4cd8-b2da-89c2407d14f2-config-volume\") pod \"collect-profiles-29399925-x74s6\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.346411 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/797e852a-cd41-4cd8-b2da-89c2407d14f2-secret-volume\") pod \"collect-profiles-29399925-x74s6\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.347490 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/797e852a-cd41-4cd8-b2da-89c2407d14f2-config-volume\") pod \"collect-profiles-29399925-x74s6\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.355432 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/797e852a-cd41-4cd8-b2da-89c2407d14f2-secret-volume\") pod \"collect-profiles-29399925-x74s6\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.373939 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llm4j\" (UniqueName: \"kubernetes.io/projected/797e852a-cd41-4cd8-b2da-89c2407d14f2-kube-api-access-llm4j\") pod \"collect-profiles-29399925-x74s6\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:00 crc kubenswrapper[4822]: I1124 14:45:00.506569 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:01 crc kubenswrapper[4822]: I1124 14:45:01.044259 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6"] Nov 24 14:45:01 crc kubenswrapper[4822]: I1124 14:45:01.811143 4822 generic.go:334] "Generic (PLEG): container finished" podID="797e852a-cd41-4cd8-b2da-89c2407d14f2" containerID="8090e6906352a4d1eef861a649b7fc57a71d5d52aa2233be3255a0571ba00677" exitCode=0 Nov 24 14:45:01 crc kubenswrapper[4822]: I1124 14:45:01.811293 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" event={"ID":"797e852a-cd41-4cd8-b2da-89c2407d14f2","Type":"ContainerDied","Data":"8090e6906352a4d1eef861a649b7fc57a71d5d52aa2233be3255a0571ba00677"} Nov 24 14:45:01 crc kubenswrapper[4822]: I1124 14:45:01.811554 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" event={"ID":"797e852a-cd41-4cd8-b2da-89c2407d14f2","Type":"ContainerStarted","Data":"0d9264bc5ca7ff228cc2d3fa33f5d632769e39c1f8319ac6217efd5ef4c65b4a"} Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.285104 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.430863 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/797e852a-cd41-4cd8-b2da-89c2407d14f2-secret-volume\") pod \"797e852a-cd41-4cd8-b2da-89c2407d14f2\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.430933 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llm4j\" (UniqueName: \"kubernetes.io/projected/797e852a-cd41-4cd8-b2da-89c2407d14f2-kube-api-access-llm4j\") pod \"797e852a-cd41-4cd8-b2da-89c2407d14f2\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.431155 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/797e852a-cd41-4cd8-b2da-89c2407d14f2-config-volume\") pod \"797e852a-cd41-4cd8-b2da-89c2407d14f2\" (UID: \"797e852a-cd41-4cd8-b2da-89c2407d14f2\") " Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.431784 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/797e852a-cd41-4cd8-b2da-89c2407d14f2-config-volume" (OuterVolumeSpecName: "config-volume") pod "797e852a-cd41-4cd8-b2da-89c2407d14f2" (UID: "797e852a-cd41-4cd8-b2da-89c2407d14f2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.436595 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797e852a-cd41-4cd8-b2da-89c2407d14f2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "797e852a-cd41-4cd8-b2da-89c2407d14f2" (UID: "797e852a-cd41-4cd8-b2da-89c2407d14f2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.436627 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797e852a-cd41-4cd8-b2da-89c2407d14f2-kube-api-access-llm4j" (OuterVolumeSpecName: "kube-api-access-llm4j") pod "797e852a-cd41-4cd8-b2da-89c2407d14f2" (UID: "797e852a-cd41-4cd8-b2da-89c2407d14f2"). InnerVolumeSpecName "kube-api-access-llm4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.534656 4822 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/797e852a-cd41-4cd8-b2da-89c2407d14f2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.534707 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llm4j\" (UniqueName: \"kubernetes.io/projected/797e852a-cd41-4cd8-b2da-89c2407d14f2-kube-api-access-llm4j\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.534730 4822 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/797e852a-cd41-4cd8-b2da-89c2407d14f2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.838920 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" event={"ID":"797e852a-cd41-4cd8-b2da-89c2407d14f2","Type":"ContainerDied","Data":"0d9264bc5ca7ff228cc2d3fa33f5d632769e39c1f8319ac6217efd5ef4c65b4a"} Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.838974 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d9264bc5ca7ff228cc2d3fa33f5d632769e39c1f8319ac6217efd5ef4c65b4a" Nov 24 14:45:03 crc kubenswrapper[4822]: I1124 14:45:03.838980 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-x74s6" Nov 24 14:45:09 crc kubenswrapper[4822]: E1124 14:45:09.265557 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" podUID="753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf" Nov 24 14:45:09 crc kubenswrapper[4822]: I1124 14:45:09.915233 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.528806 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:45:11 crc kubenswrapper[4822]: E1124 14:45:11.529122 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:45:11 crc kubenswrapper[4822]: E1124 14:45:11.529444 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:45:11 crc kubenswrapper[4822]: E1124 14:45:11.529527 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:47:13.529499391 +0000 UTC m=+1670.646139908 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.578665 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.578761 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.578844 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.580368 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.580534 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" gracePeriod=600 Nov 24 14:45:11 crc kubenswrapper[4822]: E1124 14:45:11.716310 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.940561 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" exitCode=0 Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.940708 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678"} Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.941126 4822 scope.go:117] "RemoveContainer" containerID="a9b7bf9be06a9bc314261db2c2d98effa00916294896221883633ba00a47f5d1" Nov 24 14:45:11 crc kubenswrapper[4822]: I1124 14:45:11.941919 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:45:11 crc kubenswrapper[4822]: E1124 14:45:11.942264 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:45:22 crc kubenswrapper[4822]: I1124 14:45:22.705061 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:45:22 crc kubenswrapper[4822]: E1124 14:45:22.706335 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:45:35 crc kubenswrapper[4822]: I1124 14:45:35.705390 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:45:35 crc kubenswrapper[4822]: E1124 14:45:35.706399 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:45:49 crc kubenswrapper[4822]: I1124 14:45:49.705309 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:45:49 crc kubenswrapper[4822]: E1124 14:45:49.706137 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:46:00 crc kubenswrapper[4822]: I1124 14:46:00.705451 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:46:00 crc kubenswrapper[4822]: E1124 14:46:00.706509 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.351490 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cf57w"] Nov 24 14:46:12 crc kubenswrapper[4822]: E1124 14:46:12.353067 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797e852a-cd41-4cd8-b2da-89c2407d14f2" containerName="collect-profiles" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.353103 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="797e852a-cd41-4cd8-b2da-89c2407d14f2" containerName="collect-profiles" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.353574 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="797e852a-cd41-4cd8-b2da-89c2407d14f2" containerName="collect-profiles" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.356494 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.385595 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cf57w"] Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.474667 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfkx8\" (UniqueName: \"kubernetes.io/projected/d10813b8-7910-4728-bc1e-b2e013988374-kube-api-access-hfkx8\") pod \"redhat-operators-cf57w\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.474811 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-catalog-content\") pod \"redhat-operators-cf57w\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.475044 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-utilities\") pod \"redhat-operators-cf57w\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.577525 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfkx8\" (UniqueName: \"kubernetes.io/projected/d10813b8-7910-4728-bc1e-b2e013988374-kube-api-access-hfkx8\") pod \"redhat-operators-cf57w\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.577668 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-catalog-content\") pod \"redhat-operators-cf57w\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.577744 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-utilities\") pod \"redhat-operators-cf57w\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.578336 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-catalog-content\") pod \"redhat-operators-cf57w\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.578383 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-utilities\") pod \"redhat-operators-cf57w\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.601784 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfkx8\" (UniqueName: \"kubernetes.io/projected/d10813b8-7910-4728-bc1e-b2e013988374-kube-api-access-hfkx8\") pod \"redhat-operators-cf57w\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:12 crc kubenswrapper[4822]: I1124 14:46:12.679900 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:13 crc kubenswrapper[4822]: I1124 14:46:13.214832 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cf57w"] Nov 24 14:46:13 crc kubenswrapper[4822]: I1124 14:46:13.788455 4822 generic.go:334] "Generic (PLEG): container finished" podID="d10813b8-7910-4728-bc1e-b2e013988374" containerID="3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c" exitCode=0 Nov 24 14:46:13 crc kubenswrapper[4822]: I1124 14:46:13.788503 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf57w" event={"ID":"d10813b8-7910-4728-bc1e-b2e013988374","Type":"ContainerDied","Data":"3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c"} Nov 24 14:46:13 crc kubenswrapper[4822]: I1124 14:46:13.788534 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf57w" event={"ID":"d10813b8-7910-4728-bc1e-b2e013988374","Type":"ContainerStarted","Data":"61d5f36609c2224d2403db6b218314411a759bd1956077dabd63563a9a77f059"} Nov 24 14:46:13 crc kubenswrapper[4822]: I1124 14:46:13.790760 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:46:14 crc kubenswrapper[4822]: I1124 14:46:14.800721 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf57w" event={"ID":"d10813b8-7910-4728-bc1e-b2e013988374","Type":"ContainerStarted","Data":"69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9"} Nov 24 14:46:15 crc kubenswrapper[4822]: I1124 14:46:15.704477 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:46:15 crc kubenswrapper[4822]: E1124 14:46:15.704819 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:46:15 crc kubenswrapper[4822]: I1124 14:46:15.812159 4822 generic.go:334] "Generic (PLEG): container finished" podID="d10813b8-7910-4728-bc1e-b2e013988374" containerID="69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9" exitCode=0 Nov 24 14:46:15 crc kubenswrapper[4822]: I1124 14:46:15.812202 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf57w" event={"ID":"d10813b8-7910-4728-bc1e-b2e013988374","Type":"ContainerDied","Data":"69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9"} Nov 24 14:46:16 crc kubenswrapper[4822]: I1124 14:46:16.825197 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf57w" event={"ID":"d10813b8-7910-4728-bc1e-b2e013988374","Type":"ContainerStarted","Data":"48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1"} Nov 24 14:46:16 crc kubenswrapper[4822]: I1124 14:46:16.860427 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cf57w" podStartSLOduration=2.333333228 podStartE2EDuration="4.860401418s" podCreationTimestamp="2025-11-24 14:46:12 +0000 UTC" firstStartedPulling="2025-11-24 14:46:13.79054107 +0000 UTC m=+1610.907181547" lastFinishedPulling="2025-11-24 14:46:16.31760925 +0000 UTC m=+1613.434249737" observedRunningTime="2025-11-24 14:46:16.848938432 +0000 UTC m=+1613.965578949" watchObservedRunningTime="2025-11-24 14:46:16.860401418 +0000 UTC m=+1613.977041925" Nov 24 14:46:22 crc kubenswrapper[4822]: I1124 14:46:22.680818 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:22 crc kubenswrapper[4822]: I1124 14:46:22.682179 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:22 crc kubenswrapper[4822]: I1124 14:46:22.749373 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:23 crc kubenswrapper[4822]: I1124 14:46:23.009905 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:23 crc kubenswrapper[4822]: I1124 14:46:23.081518 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cf57w"] Nov 24 14:46:24 crc kubenswrapper[4822]: I1124 14:46:24.943538 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cf57w" podUID="d10813b8-7910-4728-bc1e-b2e013988374" containerName="registry-server" containerID="cri-o://48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1" gracePeriod=2 Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.588736 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.700274 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-utilities\") pod \"d10813b8-7910-4728-bc1e-b2e013988374\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.700781 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-catalog-content\") pod \"d10813b8-7910-4728-bc1e-b2e013988374\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.701175 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfkx8\" (UniqueName: \"kubernetes.io/projected/d10813b8-7910-4728-bc1e-b2e013988374-kube-api-access-hfkx8\") pod \"d10813b8-7910-4728-bc1e-b2e013988374\" (UID: \"d10813b8-7910-4728-bc1e-b2e013988374\") " Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.701577 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-utilities" (OuterVolumeSpecName: "utilities") pod "d10813b8-7910-4728-bc1e-b2e013988374" (UID: "d10813b8-7910-4728-bc1e-b2e013988374"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.702618 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.708034 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d10813b8-7910-4728-bc1e-b2e013988374-kube-api-access-hfkx8" (OuterVolumeSpecName: "kube-api-access-hfkx8") pod "d10813b8-7910-4728-bc1e-b2e013988374" (UID: "d10813b8-7910-4728-bc1e-b2e013988374"). InnerVolumeSpecName "kube-api-access-hfkx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.809340 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfkx8\" (UniqueName: \"kubernetes.io/projected/d10813b8-7910-4728-bc1e-b2e013988374-kube-api-access-hfkx8\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.851320 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d10813b8-7910-4728-bc1e-b2e013988374" (UID: "d10813b8-7910-4728-bc1e-b2e013988374"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.910950 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d10813b8-7910-4728-bc1e-b2e013988374-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.966523 4822 generic.go:334] "Generic (PLEG): container finished" podID="d10813b8-7910-4728-bc1e-b2e013988374" containerID="48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1" exitCode=0 Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.966603 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf57w" event={"ID":"d10813b8-7910-4728-bc1e-b2e013988374","Type":"ContainerDied","Data":"48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1"} Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.966682 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf57w" event={"ID":"d10813b8-7910-4728-bc1e-b2e013988374","Type":"ContainerDied","Data":"61d5f36609c2224d2403db6b218314411a759bd1956077dabd63563a9a77f059"} Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.966707 4822 scope.go:117] "RemoveContainer" containerID="48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1" Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.968021 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf57w" Nov 24 14:46:25 crc kubenswrapper[4822]: I1124 14:46:25.989093 4822 scope.go:117] "RemoveContainer" containerID="69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.014714 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cf57w"] Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.022607 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cf57w"] Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.023411 4822 scope.go:117] "RemoveContainer" containerID="3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.078858 4822 scope.go:117] "RemoveContainer" containerID="48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1" Nov 24 14:46:26 crc kubenswrapper[4822]: E1124 14:46:26.079657 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1\": container with ID starting with 48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1 not found: ID does not exist" containerID="48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.079725 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1"} err="failed to get container status \"48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1\": rpc error: code = NotFound desc = could not find container \"48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1\": container with ID starting with 48d844bcdb61c6e2bf25a75186c57667fa535d9b9da1f8594eb4580f65c505a1 not found: ID does not exist" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.079769 4822 scope.go:117] "RemoveContainer" containerID="69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9" Nov 24 14:46:26 crc kubenswrapper[4822]: E1124 14:46:26.080368 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9\": container with ID starting with 69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9 not found: ID does not exist" containerID="69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.080465 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9"} err="failed to get container status \"69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9\": rpc error: code = NotFound desc = could not find container \"69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9\": container with ID starting with 69119a1958c9f055b581a6e54ed4664c6bab99cb215c0e6b0e4b13d6e77482c9 not found: ID does not exist" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.080606 4822 scope.go:117] "RemoveContainer" containerID="3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c" Nov 24 14:46:26 crc kubenswrapper[4822]: E1124 14:46:26.081298 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c\": container with ID starting with 3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c not found: ID does not exist" containerID="3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.081507 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c"} err="failed to get container status \"3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c\": rpc error: code = NotFound desc = could not find container \"3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c\": container with ID starting with 3168dbbc72c5b59592078045405bd372df3ff584b051903a1554f0532eb38d9c not found: ID does not exist" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.231390 4822 scope.go:117] "RemoveContainer" containerID="e6533dec0b4a77a1cc2e167c2758c732b27d3561356b0dd94e42324fb97a8c32" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.270037 4822 scope.go:117] "RemoveContainer" containerID="68203306cc053545db28dc12b820d204bf4b6b3d17b43211903ae7540f24bfdb" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.319882 4822 scope.go:117] "RemoveContainer" containerID="d9ce46ddce6957a4cbccfd3b29b2280a7a8c5722240266bb48cc308f76559e11" Nov 24 14:46:26 crc kubenswrapper[4822]: I1124 14:46:26.350288 4822 scope.go:117] "RemoveContainer" containerID="63b708115911fa584ad96a7bf285d6763f12a71f178e4989618107f79a33305d" Nov 24 14:46:27 crc kubenswrapper[4822]: I1124 14:46:27.726160 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d10813b8-7910-4728-bc1e-b2e013988374" path="/var/lib/kubelet/pods/d10813b8-7910-4728-bc1e-b2e013988374/volumes" Nov 24 14:46:28 crc kubenswrapper[4822]: I1124 14:46:28.704290 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:46:28 crc kubenswrapper[4822]: E1124 14:46:28.704948 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:46:40 crc kubenswrapper[4822]: I1124 14:46:40.705794 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:46:40 crc kubenswrapper[4822]: E1124 14:46:40.706598 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:46:48 crc kubenswrapper[4822]: I1124 14:46:48.096212 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7615-account-create-wzxch"] Nov 24 14:46:48 crc kubenswrapper[4822]: I1124 14:46:48.108281 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-tmfgk"] Nov 24 14:46:48 crc kubenswrapper[4822]: I1124 14:46:48.121107 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-qwctq"] Nov 24 14:46:48 crc kubenswrapper[4822]: I1124 14:46:48.132845 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-bfd9-account-create-khpb9"] Nov 24 14:46:48 crc kubenswrapper[4822]: I1124 14:46:48.141794 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-tmfgk"] Nov 24 14:46:48 crc kubenswrapper[4822]: I1124 14:46:48.149964 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7615-account-create-wzxch"] Nov 24 14:46:48 crc kubenswrapper[4822]: I1124 14:46:48.159875 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-bfd9-account-create-khpb9"] Nov 24 14:46:48 crc kubenswrapper[4822]: I1124 14:46:48.168089 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-qwctq"] Nov 24 14:46:49 crc kubenswrapper[4822]: I1124 14:46:49.716529 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="103881b0-811b-4c80-9ca4-d6a105da608f" path="/var/lib/kubelet/pods/103881b0-811b-4c80-9ca4-d6a105da608f/volumes" Nov 24 14:46:49 crc kubenswrapper[4822]: I1124 14:46:49.717628 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6" path="/var/lib/kubelet/pods/2569eca8-4bd8-4fcb-a9dd-c710ee60c6a6/volumes" Nov 24 14:46:49 crc kubenswrapper[4822]: I1124 14:46:49.718555 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e3ccadd-2328-41b6-88d8-b833f08073f8" path="/var/lib/kubelet/pods/2e3ccadd-2328-41b6-88d8-b833f08073f8/volumes" Nov 24 14:46:49 crc kubenswrapper[4822]: I1124 14:46:49.719384 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df77767f-08e4-4292-94b1-98d55432c038" path="/var/lib/kubelet/pods/df77767f-08e4-4292-94b1-98d55432c038/volumes" Nov 24 14:46:51 crc kubenswrapper[4822]: I1124 14:46:51.704742 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:46:51 crc kubenswrapper[4822]: E1124 14:46:51.705399 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:46:53 crc kubenswrapper[4822]: I1124 14:46:53.233136 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:46:53 crc kubenswrapper[4822]: E1124 14:46:53.233431 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:46:53 crc kubenswrapper[4822]: E1124 14:46:53.234578 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:46:53 crc kubenswrapper[4822]: E1124 14:46:53.234680 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:48:55.234647225 +0000 UTC m=+1772.351287732 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:46:54 crc kubenswrapper[4822]: E1124 14:46:54.672727 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="4acc7e6a-472b-468a-b709-183f8b3c2b5b" Nov 24 14:46:55 crc kubenswrapper[4822]: I1124 14:46:55.363687 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:46:57 crc kubenswrapper[4822]: I1124 14:46:57.027010 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-lt4wx"] Nov 24 14:46:57 crc kubenswrapper[4822]: I1124 14:46:57.036339 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-72ec-account-create-hgstv"] Nov 24 14:46:57 crc kubenswrapper[4822]: I1124 14:46:57.044578 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-lt4wx"] Nov 24 14:46:57 crc kubenswrapper[4822]: I1124 14:46:57.054234 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-72ec-account-create-hgstv"] Nov 24 14:46:57 crc kubenswrapper[4822]: I1124 14:46:57.725555 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53ec312d-2011-41f0-8a13-4268f81bfaa6" path="/var/lib/kubelet/pods/53ec312d-2011-41f0-8a13-4268f81bfaa6/volumes" Nov 24 14:46:57 crc kubenswrapper[4822]: I1124 14:46:57.727614 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67" path="/var/lib/kubelet/pods/ccc360f4-dd38-4fdc-a0e2-b9c6d3697f67/volumes" Nov 24 14:47:06 crc kubenswrapper[4822]: I1124 14:47:06.705027 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:47:06 crc kubenswrapper[4822]: E1124 14:47:06.706143 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:47:12 crc kubenswrapper[4822]: E1124 14:47:12.917425 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" podUID="753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf" Nov 24 14:47:13 crc kubenswrapper[4822]: I1124 14:47:13.530623 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:47:13 crc kubenswrapper[4822]: E1124 14:47:13.530872 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:47:13 crc kubenswrapper[4822]: E1124 14:47:13.531239 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:47:13 crc kubenswrapper[4822]: E1124 14:47:13.531354 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:49:15.53131812 +0000 UTC m=+1792.647958637 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:47:13 crc kubenswrapper[4822]: I1124 14:47:13.680053 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:47:14 crc kubenswrapper[4822]: I1124 14:47:14.046079 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-bwmcx"] Nov 24 14:47:14 crc kubenswrapper[4822]: I1124 14:47:14.063767 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-bwmcx"] Nov 24 14:47:15 crc kubenswrapper[4822]: I1124 14:47:15.726781 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a39cbd81-d3b4-46a5-a513-031aca8e43c3" path="/var/lib/kubelet/pods/a39cbd81-d3b4-46a5-a513-031aca8e43c3/volumes" Nov 24 14:47:17 crc kubenswrapper[4822]: I1124 14:47:17.705323 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:47:17 crc kubenswrapper[4822]: E1124 14:47:17.705929 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.460421 4822 scope.go:117] "RemoveContainer" containerID="2a2ffcf65260a0c02d185bd40db158f206ffa6799b48260a4fd54f5cfd59d59d" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.487638 4822 scope.go:117] "RemoveContainer" containerID="6f803271be52efcd62812b152e9ed55bc4a83f39c3abf960780134527d6744a9" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.547842 4822 scope.go:117] "RemoveContainer" containerID="1d48777a560e08f1d631ef9c9bb5c89fa5c4a6ebce7b125a6b75ae8e7249b12d" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.590343 4822 scope.go:117] "RemoveContainer" containerID="89511df05f18c959e4d41f1fab6a60626eb20e9dde051f4901b84a8e8f6d32c3" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.628103 4822 scope.go:117] "RemoveContainer" containerID="632197d2fbb1062cdb8cead4c3f4fda86897711b08e9c85541fcfcca3e1c3d09" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.674047 4822 scope.go:117] "RemoveContainer" containerID="150e93e327aae913416b5879fd99b2216c87224bb18ee1536a304ba99bb91e94" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.707831 4822 scope.go:117] "RemoveContainer" containerID="8eae7a022b1425e37963866e5fb38f9f9138c0366bfe83c9fa3e25e457ae6120" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.730925 4822 scope.go:117] "RemoveContainer" containerID="046279f7bbddb28e4e582ee2dde1a05edf5e7e1d995cab69ba63764c65fbde1c" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.783879 4822 scope.go:117] "RemoveContainer" containerID="ed54bf9ba2cb8556e6e58da64d10955d5d8ee390029607516d8b275f5cd4c346" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.801712 4822 scope.go:117] "RemoveContainer" containerID="5cdd7312b8f31f4da668e8435e18cb21763c4c4b728a4511363de44b15fcdf53" Nov 24 14:47:26 crc kubenswrapper[4822]: I1124 14:47:26.822761 4822 scope.go:117] "RemoveContainer" containerID="9f74a7941b7027868abcbfbf89668d256b3f8a261913d05d9a08810b9b88fedc" Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.076964 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-83e5-account-create-ttzgw"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.096621 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-3aed-account-create-db4jq"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.110619 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-lhp86"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.119231 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5f21-account-create-ds2h6"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.129490 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-create-ltdt2"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.137790 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8fc4-account-create-tgmfq"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.145988 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-bhnlm"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.155029 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-83e5-account-create-ttzgw"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.162947 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-4kdxz"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.170758 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-3aed-account-create-db4jq"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.178560 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-lhp86"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.185990 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-bhnlm"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.193191 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-create-ltdt2"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.200881 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8fc4-account-create-tgmfq"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.208688 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5f21-account-create-ds2h6"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.216974 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-4kdxz"] Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.726176 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fc8a762-b8f7-4246-aaec-d6670cf715a5" path="/var/lib/kubelet/pods/2fc8a762-b8f7-4246-aaec-d6670cf715a5/volumes" Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.728472 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32b5e812-c653-4437-ae27-b7a1635085a6" path="/var/lib/kubelet/pods/32b5e812-c653-4437-ae27-b7a1635085a6/volumes" Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.730042 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5254b05c-78e8-49fa-932b-4f36b5fb907d" path="/var/lib/kubelet/pods/5254b05c-78e8-49fa-932b-4f36b5fb907d/volumes" Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.732510 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="728c9997-66cf-4996-8b4d-75af0e3ba55c" path="/var/lib/kubelet/pods/728c9997-66cf-4996-8b4d-75af0e3ba55c/volumes" Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.734644 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="908b60c2-798e-4feb-90da-4cd08d479578" path="/var/lib/kubelet/pods/908b60c2-798e-4feb-90da-4cd08d479578/volumes" Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.735408 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c97044-4916-4bdd-aae7-33069d43663e" path="/var/lib/kubelet/pods/a1c97044-4916-4bdd-aae7-33069d43663e/volumes" Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.736108 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc04d3ff-c5b1-4134-b544-ca9341af72b4" path="/var/lib/kubelet/pods/bc04d3ff-c5b1-4134-b544-ca9341af72b4/volumes" Nov 24 14:47:29 crc kubenswrapper[4822]: I1124 14:47:29.737566 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cae75d4f-b139-4b12-b581-3362d46de89e" path="/var/lib/kubelet/pods/cae75d4f-b139-4b12-b581-3362d46de89e/volumes" Nov 24 14:47:32 crc kubenswrapper[4822]: I1124 14:47:32.706071 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:47:32 crc kubenswrapper[4822]: E1124 14:47:32.707004 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:47:33 crc kubenswrapper[4822]: I1124 14:47:33.047814 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-57687"] Nov 24 14:47:33 crc kubenswrapper[4822]: I1124 14:47:33.064196 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-57687"] Nov 24 14:47:33 crc kubenswrapper[4822]: I1124 14:47:33.724591 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="274d41ac-17fe-48ef-ad92-f769c96b219b" path="/var/lib/kubelet/pods/274d41ac-17fe-48ef-ad92-f769c96b219b/volumes" Nov 24 14:47:46 crc kubenswrapper[4822]: I1124 14:47:46.705786 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:47:46 crc kubenswrapper[4822]: E1124 14:47:46.707160 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:47:59 crc kubenswrapper[4822]: I1124 14:47:59.704164 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:47:59 crc kubenswrapper[4822]: E1124 14:47:59.704931 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:48:02 crc kubenswrapper[4822]: I1124 14:48:02.050108 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-8rtpz"] Nov 24 14:48:02 crc kubenswrapper[4822]: I1124 14:48:02.063987 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-8rtpz"] Nov 24 14:48:03 crc kubenswrapper[4822]: I1124 14:48:03.733500 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35fdf286-3217-480f-9522-3d146be7ceab" path="/var/lib/kubelet/pods/35fdf286-3217-480f-9522-3d146be7ceab/volumes" Nov 24 14:48:11 crc kubenswrapper[4822]: I1124 14:48:11.704834 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:48:11 crc kubenswrapper[4822]: E1124 14:48:11.705958 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:48:19 crc kubenswrapper[4822]: I1124 14:48:19.089573 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-fxf96"] Nov 24 14:48:19 crc kubenswrapper[4822]: I1124 14:48:19.107523 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-7q6rz"] Nov 24 14:48:19 crc kubenswrapper[4822]: I1124 14:48:19.125371 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-227wq"] Nov 24 14:48:19 crc kubenswrapper[4822]: I1124 14:48:19.132322 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-fxf96"] Nov 24 14:48:19 crc kubenswrapper[4822]: I1124 14:48:19.139618 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-7q6rz"] Nov 24 14:48:19 crc kubenswrapper[4822]: I1124 14:48:19.147013 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-227wq"] Nov 24 14:48:19 crc kubenswrapper[4822]: I1124 14:48:19.721410 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42faa16a-d5df-4002-94bc-dd9b727ad202" path="/var/lib/kubelet/pods/42faa16a-d5df-4002-94bc-dd9b727ad202/volumes" Nov 24 14:48:19 crc kubenswrapper[4822]: I1124 14:48:19.723096 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cebbef1-88ea-4dd3-9a1d-3900d2d160bd" path="/var/lib/kubelet/pods/5cebbef1-88ea-4dd3-9a1d-3900d2d160bd/volumes" Nov 24 14:48:19 crc kubenswrapper[4822]: I1124 14:48:19.724276 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="858eb7c9-030e-49d8-b018-fdf41e979d20" path="/var/lib/kubelet/pods/858eb7c9-030e-49d8-b018-fdf41e979d20/volumes" Nov 24 14:48:22 crc kubenswrapper[4822]: I1124 14:48:22.704501 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:48:22 crc kubenswrapper[4822]: E1124 14:48:22.704912 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.005825 4822 scope.go:117] "RemoveContainer" containerID="06c06097721e5f3337bcd2702a2f65c9e5e376dfef37c33d0b09c83d8085e70a" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.094508 4822 scope.go:117] "RemoveContainer" containerID="60b15c160bd848838dc3599ba9bad2bc7838cc8b2ebfd95c2721f7dca97fb6ff" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.144009 4822 scope.go:117] "RemoveContainer" containerID="3d2c9ab3c17bdf612aae63973f64a30d64aefbdb503d7e6ad009200153c7bf7e" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.180394 4822 scope.go:117] "RemoveContainer" containerID="e94a0fac2144c591a1d37f88873d350885e2f13e7467396e159a9984f10a8c05" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.223135 4822 scope.go:117] "RemoveContainer" containerID="e7e61c4499846d98aed167fd29eaebb8746aded78421eb4f90753c6cc30bf6fd" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.289184 4822 scope.go:117] "RemoveContainer" containerID="9fee266f7490312ca4eb42305221c99f63fc4fc82744130d03bc660f88f60949" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.334273 4822 scope.go:117] "RemoveContainer" containerID="5ef5cc9c7238acea58a84c5a18958f784a38b03e541d20ea658682453ba33a37" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.402042 4822 scope.go:117] "RemoveContainer" containerID="91b1804b1e534f9081e6b584352489cfb8d984eef5a0764bc62dffe333e3bacd" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.446157 4822 scope.go:117] "RemoveContainer" containerID="ead076bf4ebebd558975431b3a838a550c7480d92a1fa7b2a68615ec0bc34101" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.471621 4822 scope.go:117] "RemoveContainer" containerID="3fa43b7ab86ca72cd1e37740e9b76218c196ac8152192ac8a570513ea72f9242" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.504494 4822 scope.go:117] "RemoveContainer" containerID="0aac29ea10e6cfe52c197751f5798769a1e2f22a4e15a51e51eba8666a586266" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.547588 4822 scope.go:117] "RemoveContainer" containerID="7e65a9dd55dc67b0e7d2c121ce972f4b2dfc73b91d356be0d73b997a699db30c" Nov 24 14:48:27 crc kubenswrapper[4822]: I1124 14:48:27.573046 4822 scope.go:117] "RemoveContainer" containerID="0fc918670c183ca24392e2580eca485a0ebecfebdd000e8956d4ff66d7f2bacb" Nov 24 14:48:31 crc kubenswrapper[4822]: I1124 14:48:31.045796 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-fzrq8"] Nov 24 14:48:31 crc kubenswrapper[4822]: I1124 14:48:31.057961 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-fzrq8"] Nov 24 14:48:31 crc kubenswrapper[4822]: I1124 14:48:31.719311 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7348b48b-1f84-439f-a049-bfc19935dbfc" path="/var/lib/kubelet/pods/7348b48b-1f84-439f-a049-bfc19935dbfc/volumes" Nov 24 14:48:32 crc kubenswrapper[4822]: I1124 14:48:32.048484 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-s6qvf"] Nov 24 14:48:32 crc kubenswrapper[4822]: I1124 14:48:32.063489 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-s6qvf"] Nov 24 14:48:33 crc kubenswrapper[4822]: I1124 14:48:33.731519 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0ad3a91-d60c-40a0-a285-c5a9615fffe0" path="/var/lib/kubelet/pods/f0ad3a91-d60c-40a0-a285-c5a9615fffe0/volumes" Nov 24 14:48:36 crc kubenswrapper[4822]: I1124 14:48:36.705674 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:48:36 crc kubenswrapper[4822]: E1124 14:48:36.706630 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:48:40 crc kubenswrapper[4822]: I1124 14:48:40.044977 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-j72lq"] Nov 24 14:48:40 crc kubenswrapper[4822]: I1124 14:48:40.056956 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-j72lq"] Nov 24 14:48:41 crc kubenswrapper[4822]: I1124 14:48:41.724011 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c25c6942-9f38-4db4-834f-2262ee899ed7" path="/var/lib/kubelet/pods/c25c6942-9f38-4db4-834f-2262ee899ed7/volumes" Nov 24 14:48:48 crc kubenswrapper[4822]: I1124 14:48:48.705045 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:48:48 crc kubenswrapper[4822]: E1124 14:48:48.706678 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:48:55 crc kubenswrapper[4822]: I1124 14:48:55.263107 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:48:55 crc kubenswrapper[4822]: E1124 14:48:55.263511 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:48:55 crc kubenswrapper[4822]: E1124 14:48:55.263808 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:48:55 crc kubenswrapper[4822]: E1124 14:48:55.263892 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift podName:4acc7e6a-472b-468a-b709-183f8b3c2b5b nodeName:}" failed. No retries permitted until 2025-11-24 14:50:57.263869107 +0000 UTC m=+1894.380509614 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift") pod "swift-storage-0" (UID: "4acc7e6a-472b-468a-b709-183f8b3c2b5b") : configmap "swift-ring-files" not found Nov 24 14:48:58 crc kubenswrapper[4822]: E1124 14:48:58.366190 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="4acc7e6a-472b-468a-b709-183f8b3c2b5b" Nov 24 14:48:59 crc kubenswrapper[4822]: I1124 14:48:59.120446 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:49:00 crc kubenswrapper[4822]: I1124 14:49:00.706268 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:49:00 crc kubenswrapper[4822]: E1124 14:49:00.707016 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:49:15 crc kubenswrapper[4822]: I1124 14:49:15.537541 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:49:15 crc kubenswrapper[4822]: E1124 14:49:15.537801 4822 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:49:15 crc kubenswrapper[4822]: E1124 14:49:15.538249 4822 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-6b8fb4f557-bkwxl: configmap "swift-ring-files" not found Nov 24 14:49:15 crc kubenswrapper[4822]: E1124 14:49:15.538407 4822 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift podName:753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf nodeName:}" failed. No retries permitted until 2025-11-24 14:51:17.538364363 +0000 UTC m=+1914.655004890 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift") pod "swift-proxy-6b8fb4f557-bkwxl" (UID: "753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf") : configmap "swift-ring-files" not found Nov 24 14:49:15 crc kubenswrapper[4822]: I1124 14:49:15.704564 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:49:15 crc kubenswrapper[4822]: E1124 14:49:15.704887 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:49:16 crc kubenswrapper[4822]: E1124 14:49:16.681800 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" podUID="753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf" Nov 24 14:49:17 crc kubenswrapper[4822]: I1124 14:49:17.307971 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:49:26 crc kubenswrapper[4822]: I1124 14:49:26.705517 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:49:26 crc kubenswrapper[4822]: E1124 14:49:26.706762 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:49:27 crc kubenswrapper[4822]: I1124 14:49:27.923545 4822 scope.go:117] "RemoveContainer" containerID="16409edc9499b818433aed5a90f6f114cc12534d16259515b0d3505fd5a7f229" Nov 24 14:49:27 crc kubenswrapper[4822]: I1124 14:49:27.958304 4822 scope.go:117] "RemoveContainer" containerID="38df6d787c8345ee1050f3c2b8490298e447001c5be7a4220ce71bc14a8df2e0" Nov 24 14:49:28 crc kubenswrapper[4822]: I1124 14:49:28.026225 4822 scope.go:117] "RemoveContainer" containerID="25897687c5bab918dbe59432ac3f4cc18ba2f64eaff93cc50236c796cd596630" Nov 24 14:49:36 crc kubenswrapper[4822]: I1124 14:49:36.062401 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c404-account-create-lpw6b"] Nov 24 14:49:36 crc kubenswrapper[4822]: I1124 14:49:36.070240 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-nhlp2"] Nov 24 14:49:36 crc kubenswrapper[4822]: I1124 14:49:36.078170 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-mr4wn"] Nov 24 14:49:36 crc kubenswrapper[4822]: I1124 14:49:36.084508 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c404-account-create-lpw6b"] Nov 24 14:49:36 crc kubenswrapper[4822]: I1124 14:49:36.091094 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-nhlp2"] Nov 24 14:49:36 crc kubenswrapper[4822]: I1124 14:49:36.097171 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-mr4wn"] Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.048252 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-5c5d-account-create-t7dqm"] Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.075726 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-bd5c-account-create-6z47s"] Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.085798 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-5c5d-account-create-t7dqm"] Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.094633 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-smgtd"] Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.102194 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-bd5c-account-create-6z47s"] Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.109798 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-smgtd"] Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.721577 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5edd476f-b73c-4f34-b905-baebf09ce800" path="/var/lib/kubelet/pods/5edd476f-b73c-4f34-b905-baebf09ce800/volumes" Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.722820 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e7fd483-39ae-40aa-b649-e3558dab68c3" path="/var/lib/kubelet/pods/6e7fd483-39ae-40aa-b649-e3558dab68c3/volumes" Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.723622 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86107af1-85fb-4fac-8864-ebdf9d290058" path="/var/lib/kubelet/pods/86107af1-85fb-4fac-8864-ebdf9d290058/volumes" Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.724515 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="866995d8-9b1e-4558-b713-955dd3e3621b" path="/var/lib/kubelet/pods/866995d8-9b1e-4558-b713-955dd3e3621b/volumes" Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.726029 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a102f476-c876-41c0-8700-bac9fbd2b95c" path="/var/lib/kubelet/pods/a102f476-c876-41c0-8700-bac9fbd2b95c/volumes" Nov 24 14:49:37 crc kubenswrapper[4822]: I1124 14:49:37.726991 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5ccd4f9-b868-476d-b08f-cc78590964b1" path="/var/lib/kubelet/pods/d5ccd4f9-b868-476d-b08f-cc78590964b1/volumes" Nov 24 14:49:41 crc kubenswrapper[4822]: I1124 14:49:41.705677 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:49:41 crc kubenswrapper[4822]: E1124 14:49:41.708238 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.466416 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wx524"] Nov 24 14:49:47 crc kubenswrapper[4822]: E1124 14:49:47.467884 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d10813b8-7910-4728-bc1e-b2e013988374" containerName="registry-server" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.467916 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d10813b8-7910-4728-bc1e-b2e013988374" containerName="registry-server" Nov 24 14:49:47 crc kubenswrapper[4822]: E1124 14:49:47.467958 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d10813b8-7910-4728-bc1e-b2e013988374" containerName="extract-utilities" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.467974 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d10813b8-7910-4728-bc1e-b2e013988374" containerName="extract-utilities" Nov 24 14:49:47 crc kubenswrapper[4822]: E1124 14:49:47.468022 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d10813b8-7910-4728-bc1e-b2e013988374" containerName="extract-content" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.468039 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d10813b8-7910-4728-bc1e-b2e013988374" containerName="extract-content" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.468496 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="d10813b8-7910-4728-bc1e-b2e013988374" containerName="registry-server" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.469802 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.472895 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.473455 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.497108 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wx524"] Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.509486 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfjzs\" (UniqueName: \"kubernetes.io/projected/65b23262-b92d-4338-8e81-7e594f48be7b-kube-api-access-bfjzs\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.509810 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-ring-data-devices\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.510059 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-combined-ca-bundle\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.510300 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-dispersionconf\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.510597 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-scripts\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.510781 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-swiftconf\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.510967 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65b23262-b92d-4338-8e81-7e594f48be7b-etc-swift\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.612410 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-combined-ca-bundle\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.612519 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-dispersionconf\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.612563 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-scripts\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.612615 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-swiftconf\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.612713 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65b23262-b92d-4338-8e81-7e594f48be7b-etc-swift\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.612958 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfjzs\" (UniqueName: \"kubernetes.io/projected/65b23262-b92d-4338-8e81-7e594f48be7b-kube-api-access-bfjzs\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.613025 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-ring-data-devices\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.614627 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65b23262-b92d-4338-8e81-7e594f48be7b-etc-swift\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.615394 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-ring-data-devices\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.616701 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-scripts\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.622445 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-dispersionconf\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.625310 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-combined-ca-bundle\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.625579 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-swiftconf\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.650673 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfjzs\" (UniqueName: \"kubernetes.io/projected/65b23262-b92d-4338-8e81-7e594f48be7b-kube-api-access-bfjzs\") pod \"swift-ring-rebalance-wx524\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.810790 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-d8mh6" Nov 24 14:49:47 crc kubenswrapper[4822]: I1124 14:49:47.818908 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:49:48 crc kubenswrapper[4822]: I1124 14:49:48.328936 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wx524"] Nov 24 14:49:48 crc kubenswrapper[4822]: I1124 14:49:48.639766 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wx524" event={"ID":"65b23262-b92d-4338-8e81-7e594f48be7b","Type":"ContainerStarted","Data":"659507c360563158d98fa8651605ff76a7fcbf700e983b7ae48a3f7add3069c5"} Nov 24 14:49:51 crc kubenswrapper[4822]: I1124 14:49:51.688471 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wx524" event={"ID":"65b23262-b92d-4338-8e81-7e594f48be7b","Type":"ContainerStarted","Data":"78c0a9f982d2e043ed5d5a4de0bb4b503302222137366b7ff0f0c990a2cc5338"} Nov 24 14:49:51 crc kubenswrapper[4822]: I1124 14:49:51.719712 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-wx524" podStartSLOduration=1.8039377509999999 podStartE2EDuration="4.719685095s" podCreationTimestamp="2025-11-24 14:49:47 +0000 UTC" firstStartedPulling="2025-11-24 14:49:48.3421026 +0000 UTC m=+1825.458743107" lastFinishedPulling="2025-11-24 14:49:51.257849974 +0000 UTC m=+1828.374490451" observedRunningTime="2025-11-24 14:49:51.709910853 +0000 UTC m=+1828.826551420" watchObservedRunningTime="2025-11-24 14:49:51.719685095 +0000 UTC m=+1828.836325602" Nov 24 14:49:55 crc kubenswrapper[4822]: I1124 14:49:55.705015 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:49:55 crc kubenswrapper[4822]: E1124 14:49:55.706095 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:49:58 crc kubenswrapper[4822]: I1124 14:49:58.820719 4822 generic.go:334] "Generic (PLEG): container finished" podID="65b23262-b92d-4338-8e81-7e594f48be7b" containerID="78c0a9f982d2e043ed5d5a4de0bb4b503302222137366b7ff0f0c990a2cc5338" exitCode=0 Nov 24 14:49:58 crc kubenswrapper[4822]: I1124 14:49:58.820794 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wx524" event={"ID":"65b23262-b92d-4338-8e81-7e594f48be7b","Type":"ContainerDied","Data":"78c0a9f982d2e043ed5d5a4de0bb4b503302222137366b7ff0f0c990a2cc5338"} Nov 24 14:49:59 crc kubenswrapper[4822]: I1124 14:49:59.057340 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7kfv9"] Nov 24 14:49:59 crc kubenswrapper[4822]: I1124 14:49:59.072867 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7kfv9"] Nov 24 14:49:59 crc kubenswrapper[4822]: I1124 14:49:59.725918 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c89b401-31c4-46b1-aa12-49e823e2cf21" path="/var/lib/kubelet/pods/8c89b401-31c4-46b1-aa12-49e823e2cf21/volumes" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.192787 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.260437 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-combined-ca-bundle\") pod \"65b23262-b92d-4338-8e81-7e594f48be7b\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.260890 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65b23262-b92d-4338-8e81-7e594f48be7b-etc-swift\") pod \"65b23262-b92d-4338-8e81-7e594f48be7b\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.261029 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfjzs\" (UniqueName: \"kubernetes.io/projected/65b23262-b92d-4338-8e81-7e594f48be7b-kube-api-access-bfjzs\") pod \"65b23262-b92d-4338-8e81-7e594f48be7b\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.261123 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-dispersionconf\") pod \"65b23262-b92d-4338-8e81-7e594f48be7b\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.261157 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-swiftconf\") pod \"65b23262-b92d-4338-8e81-7e594f48be7b\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.261395 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-ring-data-devices\") pod \"65b23262-b92d-4338-8e81-7e594f48be7b\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.261427 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-scripts\") pod \"65b23262-b92d-4338-8e81-7e594f48be7b\" (UID: \"65b23262-b92d-4338-8e81-7e594f48be7b\") " Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.261753 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b23262-b92d-4338-8e81-7e594f48be7b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "65b23262-b92d-4338-8e81-7e594f48be7b" (UID: "65b23262-b92d-4338-8e81-7e594f48be7b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.261950 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "65b23262-b92d-4338-8e81-7e594f48be7b" (UID: "65b23262-b92d-4338-8e81-7e594f48be7b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.262432 4822 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.262458 4822 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65b23262-b92d-4338-8e81-7e594f48be7b-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.266812 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b23262-b92d-4338-8e81-7e594f48be7b-kube-api-access-bfjzs" (OuterVolumeSpecName: "kube-api-access-bfjzs") pod "65b23262-b92d-4338-8e81-7e594f48be7b" (UID: "65b23262-b92d-4338-8e81-7e594f48be7b"). InnerVolumeSpecName "kube-api-access-bfjzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.290509 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "65b23262-b92d-4338-8e81-7e594f48be7b" (UID: "65b23262-b92d-4338-8e81-7e594f48be7b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.296329 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "65b23262-b92d-4338-8e81-7e594f48be7b" (UID: "65b23262-b92d-4338-8e81-7e594f48be7b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.305849 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-scripts" (OuterVolumeSpecName: "scripts") pod "65b23262-b92d-4338-8e81-7e594f48be7b" (UID: "65b23262-b92d-4338-8e81-7e594f48be7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.312045 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65b23262-b92d-4338-8e81-7e594f48be7b" (UID: "65b23262-b92d-4338-8e81-7e594f48be7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.364723 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65b23262-b92d-4338-8e81-7e594f48be7b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.364777 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.364804 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfjzs\" (UniqueName: \"kubernetes.io/projected/65b23262-b92d-4338-8e81-7e594f48be7b-kube-api-access-bfjzs\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.364830 4822 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.364855 4822 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65b23262-b92d-4338-8e81-7e594f48be7b-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.854771 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wx524" event={"ID":"65b23262-b92d-4338-8e81-7e594f48be7b","Type":"ContainerDied","Data":"659507c360563158d98fa8651605ff76a7fcbf700e983b7ae48a3f7add3069c5"} Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.854818 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="659507c360563158d98fa8651605ff76a7fcbf700e983b7ae48a3f7add3069c5" Nov 24 14:50:00 crc kubenswrapper[4822]: I1124 14:50:00.855381 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wx524" Nov 24 14:50:10 crc kubenswrapper[4822]: I1124 14:50:10.704924 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:50:10 crc kubenswrapper[4822]: E1124 14:50:10.705959 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:50:22 crc kubenswrapper[4822]: I1124 14:50:22.072070 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-8p47j"] Nov 24 14:50:22 crc kubenswrapper[4822]: I1124 14:50:22.092372 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-8p47j"] Nov 24 14:50:23 crc kubenswrapper[4822]: I1124 14:50:23.052119 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z2dll"] Nov 24 14:50:23 crc kubenswrapper[4822]: I1124 14:50:23.066561 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z2dll"] Nov 24 14:50:23 crc kubenswrapper[4822]: I1124 14:50:23.726228 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b7c7bb8-bcd0-4da3-aac0-204a07f26617" path="/var/lib/kubelet/pods/8b7c7bb8-bcd0-4da3-aac0-204a07f26617/volumes" Nov 24 14:50:23 crc kubenswrapper[4822]: I1124 14:50:23.727939 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdf7a5a9-222a-4012-be03-5135f9bc0b1c" path="/var/lib/kubelet/pods/bdf7a5a9-222a-4012-be03-5135f9bc0b1c/volumes" Nov 24 14:50:24 crc kubenswrapper[4822]: I1124 14:50:24.704920 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:50:25 crc kubenswrapper[4822]: I1124 14:50:25.180600 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"5cc0b1343a16a8d188981e3c048d5b5d39678fa534e29571730f78b864e29cf8"} Nov 24 14:50:28 crc kubenswrapper[4822]: I1124 14:50:28.145016 4822 scope.go:117] "RemoveContainer" containerID="07d9bdb91d4ba52bb4e67aed971b71b5a2f155b692bcdef13a2832a4efafdbce" Nov 24 14:50:28 crc kubenswrapper[4822]: I1124 14:50:28.188746 4822 scope.go:117] "RemoveContainer" containerID="d526f181dd729f0b676df402b284fef762847ea69c04ff7454d970c587fea0cc" Nov 24 14:50:28 crc kubenswrapper[4822]: I1124 14:50:28.257544 4822 scope.go:117] "RemoveContainer" containerID="1fd7145c53cea23eb0f13026d0b0427e31a0dc8964ece07e829785e5a862f6e1" Nov 24 14:50:28 crc kubenswrapper[4822]: I1124 14:50:28.312410 4822 scope.go:117] "RemoveContainer" containerID="2f267ad20a6e8b56fd209664fa21f4c243b40fba5cfae1eea9b88b8d08ffd871" Nov 24 14:50:28 crc kubenswrapper[4822]: I1124 14:50:28.373349 4822 scope.go:117] "RemoveContainer" containerID="71994cbbfdd0df3867c8c71fd5f7c9461f2e2eb0d7357f6143be0b0c4fc71675" Nov 24 14:50:28 crc kubenswrapper[4822]: I1124 14:50:28.404731 4822 scope.go:117] "RemoveContainer" containerID="e4b885cd4ea462207162e2986589f5ca3e09df63d17cec2252b6d012edec94f5" Nov 24 14:50:28 crc kubenswrapper[4822]: I1124 14:50:28.446267 4822 scope.go:117] "RemoveContainer" containerID="7b2a840d52fadd5f75579043238b07d839f79f0535a77d8b8c038d0af0b9770e" Nov 24 14:50:28 crc kubenswrapper[4822]: I1124 14:50:28.496142 4822 scope.go:117] "RemoveContainer" containerID="84cdfdc69581ec29622a5fc153a5198c8cecf84d476e10580993bf03c047c7f3" Nov 24 14:50:28 crc kubenswrapper[4822]: I1124 14:50:28.536260 4822 scope.go:117] "RemoveContainer" containerID="1753807851ea24920b8697e4b5274fdcf13f4f8476bc382c3e60a3279ae0f726" Nov 24 14:50:57 crc kubenswrapper[4822]: I1124 14:50:57.273753 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:50:57 crc kubenswrapper[4822]: I1124 14:50:57.282046 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4acc7e6a-472b-468a-b709-183f8b3c2b5b-etc-swift\") pod \"swift-storage-0\" (UID: \"4acc7e6a-472b-468a-b709-183f8b3c2b5b\") " pod="openstack/swift-storage-0" Nov 24 14:50:57 crc kubenswrapper[4822]: I1124 14:50:57.322273 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:50:57 crc kubenswrapper[4822]: I1124 14:50:57.911689 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 14:50:58 crc kubenswrapper[4822]: I1124 14:50:58.574270 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"ace963d94c5c18edaba884e28379d8bd319bd1a9c3e04f5897764f2d05047946"} Nov 24 14:50:59 crc kubenswrapper[4822]: I1124 14:50:59.603189 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"269d85834a9ea1077b77caf75f50dfa66221ae243ffed28b2c6e20a880626195"} Nov 24 14:51:00 crc kubenswrapper[4822]: I1124 14:51:00.618796 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"117e0208db122beb831c22c96938dd9a990d67098e92457ab0d3850ee6477398"} Nov 24 14:51:00 crc kubenswrapper[4822]: I1124 14:51:00.619238 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"19f5fdca6aeaf4ab6bb0498085980d02208ebc2667201689287dbd067104a527"} Nov 24 14:51:00 crc kubenswrapper[4822]: I1124 14:51:00.619248 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"5b929df835098d3d53c59c403c773c97977b9b317e7df7c8dfa82ee9e96e0ca1"} Nov 24 14:51:01 crc kubenswrapper[4822]: I1124 14:51:01.630125 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"004c6fa13b634b241feccc3d8ca841d740774107ba839e3e246ad8d561b9b7ad"} Nov 24 14:51:02 crc kubenswrapper[4822]: I1124 14:51:02.644974 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"059ee7552ec9addc583ec3ceb3c81a13ff3a4996f8e47b407b72c0f64eab9176"} Nov 24 14:51:02 crc kubenswrapper[4822]: I1124 14:51:02.645257 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"ea2c4d1821d58ca8944fbab24f540fd4b41774b27affd60529951f2ad6b3386b"} Nov 24 14:51:02 crc kubenswrapper[4822]: I1124 14:51:02.645268 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"2f054fc8714b4758f4bdc981b0cdfef1bc8c9794174f49d357e11a575ab3ea01"} Nov 24 14:51:03 crc kubenswrapper[4822]: I1124 14:51:03.671458 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"da605f3730d02ca36225bd269da028a13bee2b90a04e0b7da31bfdd7657e622d"} Nov 24 14:51:04 crc kubenswrapper[4822]: I1124 14:51:04.691881 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"0eff2bf9a584ce785e89c23c57e9cb3b7a6c0e6db96b16d36bfe8b84effd9be6"} Nov 24 14:51:04 crc kubenswrapper[4822]: I1124 14:51:04.692146 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"c70268ea6290e0673952243f5b0ab973805662751ef27b6da8beb4cecd60e9f0"} Nov 24 14:51:04 crc kubenswrapper[4822]: I1124 14:51:04.692155 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"2ea4ba99579a90c1f89bffba23e1696398a4cd0e367a8b672cad14d30c273dc9"} Nov 24 14:51:04 crc kubenswrapper[4822]: I1124 14:51:04.692162 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"04924ff80997d7ca8c19a99f33e71a78b7640e7cf2d9e71f5b98f799d8d01b3a"} Nov 24 14:51:04 crc kubenswrapper[4822]: I1124 14:51:04.692170 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"530c510cdee74c718f6deb1a876afffd96b9253d30f3ff6ef2f56fff35750078"} Nov 24 14:51:05 crc kubenswrapper[4822]: I1124 14:51:05.734802 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4acc7e6a-472b-468a-b709-183f8b3c2b5b","Type":"ContainerStarted","Data":"7e34ccb7cf956ac1f5d5095de2d7c1a0fb3934bca2d7c64809961bff85780606"} Nov 24 14:51:05 crc kubenswrapper[4822]: I1124 14:51:05.794015 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=865.376662483 podStartE2EDuration="14m30.793985495s" podCreationTimestamp="2025-11-24 14:36:35 +0000 UTC" firstStartedPulling="2025-11-24 14:50:57.918626379 +0000 UTC m=+1895.035266856" lastFinishedPulling="2025-11-24 14:51:03.335949391 +0000 UTC m=+1900.452589868" observedRunningTime="2025-11-24 14:51:05.769335774 +0000 UTC m=+1902.885976291" watchObservedRunningTime="2025-11-24 14:51:05.793985495 +0000 UTC m=+1902.910625982" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.054498 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-sxqwl"] Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.068797 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-sxqwl"] Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.193656 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bf74d94c7-kp4rh"] Nov 24 14:51:06 crc kubenswrapper[4822]: E1124 14:51:06.194187 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b23262-b92d-4338-8e81-7e594f48be7b" containerName="swift-ring-rebalance" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.194232 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b23262-b92d-4338-8e81-7e594f48be7b" containerName="swift-ring-rebalance" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.194522 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b23262-b92d-4338-8e81-7e594f48be7b" containerName="swift-ring-rebalance" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.196084 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.206915 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.207914 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bf74d94c7-kp4rh"] Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.383848 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzqft\" (UniqueName: \"kubernetes.io/projected/3d3e4572-1875-459c-9fbd-53ea02e8e123-kube-api-access-gzqft\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.383905 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-sb\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.384089 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-config\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.384142 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-nb\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.384185 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-svc\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.384243 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-swift-storage-0\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.486693 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzqft\" (UniqueName: \"kubernetes.io/projected/3d3e4572-1875-459c-9fbd-53ea02e8e123-kube-api-access-gzqft\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.486745 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-sb\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.486839 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-config\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.486873 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-nb\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.486891 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-svc\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.486913 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-swift-storage-0\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.490879 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-sb\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.490890 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-swift-storage-0\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.494781 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-svc\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.496537 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-nb\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.496649 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-config\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.528958 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzqft\" (UniqueName: \"kubernetes.io/projected/3d3e4572-1875-459c-9fbd-53ea02e8e123-kube-api-access-gzqft\") pod \"dnsmasq-dns-bf74d94c7-kp4rh\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:06 crc kubenswrapper[4822]: I1124 14:51:06.816046 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:07 crc kubenswrapper[4822]: I1124 14:51:07.312715 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bf74d94c7-kp4rh"] Nov 24 14:51:07 crc kubenswrapper[4822]: W1124 14:51:07.324288 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d3e4572_1875_459c_9fbd_53ea02e8e123.slice/crio-c6e709c36f31e650c64f829dbb3f898f0ea02428e630f822943f34b017906f42 WatchSource:0}: Error finding container c6e709c36f31e650c64f829dbb3f898f0ea02428e630f822943f34b017906f42: Status 404 returned error can't find the container with id c6e709c36f31e650c64f829dbb3f898f0ea02428e630f822943f34b017906f42 Nov 24 14:51:07 crc kubenswrapper[4822]: I1124 14:51:07.719893 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75e3bf9c-9f8b-4713-ba67-3e21635c4521" path="/var/lib/kubelet/pods/75e3bf9c-9f8b-4713-ba67-3e21635c4521/volumes" Nov 24 14:51:07 crc kubenswrapper[4822]: I1124 14:51:07.731624 4822 generic.go:334] "Generic (PLEG): container finished" podID="3d3e4572-1875-459c-9fbd-53ea02e8e123" containerID="6d8d8b424df4d1487c6a9f1bc2c3698433cee66cfe8301fd0e9184786fba1640" exitCode=0 Nov 24 14:51:07 crc kubenswrapper[4822]: I1124 14:51:07.731684 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" event={"ID":"3d3e4572-1875-459c-9fbd-53ea02e8e123","Type":"ContainerDied","Data":"6d8d8b424df4d1487c6a9f1bc2c3698433cee66cfe8301fd0e9184786fba1640"} Nov 24 14:51:07 crc kubenswrapper[4822]: I1124 14:51:07.731742 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" event={"ID":"3d3e4572-1875-459c-9fbd-53ea02e8e123","Type":"ContainerStarted","Data":"c6e709c36f31e650c64f829dbb3f898f0ea02428e630f822943f34b017906f42"} Nov 24 14:51:08 crc kubenswrapper[4822]: I1124 14:51:08.745155 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" event={"ID":"3d3e4572-1875-459c-9fbd-53ea02e8e123","Type":"ContainerStarted","Data":"e8868dda0ec2298907ee7e1a61984e27b8adf12cdc09cf103e325220b3df05ab"} Nov 24 14:51:08 crc kubenswrapper[4822]: I1124 14:51:08.745626 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:08 crc kubenswrapper[4822]: I1124 14:51:08.771399 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" podStartSLOduration=2.771379136 podStartE2EDuration="2.771379136s" podCreationTimestamp="2025-11-24 14:51:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:51:08.764991404 +0000 UTC m=+1905.881631901" watchObservedRunningTime="2025-11-24 14:51:08.771379136 +0000 UTC m=+1905.888019613" Nov 24 14:51:16 crc kubenswrapper[4822]: I1124 14:51:16.817434 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:51:16 crc kubenswrapper[4822]: I1124 14:51:16.954505 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cbd4487cc-66fhk"] Nov 24 14:51:16 crc kubenswrapper[4822]: I1124 14:51:16.954942 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" podUID="02fe13aa-aece-4235-b424-e244d720b40f" containerName="dnsmasq-dns" containerID="cri-o://aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082" gracePeriod=10 Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.506961 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.555570 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.565772 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf-etc-swift\") pod \"swift-proxy-6b8fb4f557-bkwxl\" (UID: \"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf\") " pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.608797 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.657082 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-config\") pod \"02fe13aa-aece-4235-b424-e244d720b40f\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.657185 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-dns-svc\") pod \"02fe13aa-aece-4235-b424-e244d720b40f\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.657328 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-nb\") pod \"02fe13aa-aece-4235-b424-e244d720b40f\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.657379 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-sb\") pod \"02fe13aa-aece-4235-b424-e244d720b40f\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.657465 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8s44\" (UniqueName: \"kubernetes.io/projected/02fe13aa-aece-4235-b424-e244d720b40f-kube-api-access-x8s44\") pod \"02fe13aa-aece-4235-b424-e244d720b40f\" (UID: \"02fe13aa-aece-4235-b424-e244d720b40f\") " Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.662317 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02fe13aa-aece-4235-b424-e244d720b40f-kube-api-access-x8s44" (OuterVolumeSpecName: "kube-api-access-x8s44") pod "02fe13aa-aece-4235-b424-e244d720b40f" (UID: "02fe13aa-aece-4235-b424-e244d720b40f"). InnerVolumeSpecName "kube-api-access-x8s44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.733258 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "02fe13aa-aece-4235-b424-e244d720b40f" (UID: "02fe13aa-aece-4235-b424-e244d720b40f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.745751 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-config" (OuterVolumeSpecName: "config") pod "02fe13aa-aece-4235-b424-e244d720b40f" (UID: "02fe13aa-aece-4235-b424-e244d720b40f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.747784 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "02fe13aa-aece-4235-b424-e244d720b40f" (UID: "02fe13aa-aece-4235-b424-e244d720b40f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.749752 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "02fe13aa-aece-4235-b424-e244d720b40f" (UID: "02fe13aa-aece-4235-b424-e244d720b40f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.759842 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.759856 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.759865 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.759875 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02fe13aa-aece-4235-b424-e244d720b40f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.759884 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8s44\" (UniqueName: \"kubernetes.io/projected/02fe13aa-aece-4235-b424-e244d720b40f-kube-api-access-x8s44\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.854937 4822 generic.go:334] "Generic (PLEG): container finished" podID="02fe13aa-aece-4235-b424-e244d720b40f" containerID="aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082" exitCode=0 Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.854990 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.855006 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" event={"ID":"02fe13aa-aece-4235-b424-e244d720b40f","Type":"ContainerDied","Data":"aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082"} Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.855408 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbd4487cc-66fhk" event={"ID":"02fe13aa-aece-4235-b424-e244d720b40f","Type":"ContainerDied","Data":"0c6ed307f8689199010e8e90d6673870ad734dc501dffb0803ab3ae3b5a19817"} Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.855433 4822 scope.go:117] "RemoveContainer" containerID="aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.876362 4822 scope.go:117] "RemoveContainer" containerID="83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.893958 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cbd4487cc-66fhk"] Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.899078 4822 scope.go:117] "RemoveContainer" containerID="aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082" Nov 24 14:51:17 crc kubenswrapper[4822]: E1124 14:51:17.899482 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082\": container with ID starting with aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082 not found: ID does not exist" containerID="aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.899512 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082"} err="failed to get container status \"aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082\": rpc error: code = NotFound desc = could not find container \"aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082\": container with ID starting with aa2ef83e89a6bbe7cdf69d860052764f8ead9e9b5300f6c8725b49d9890b0082 not found: ID does not exist" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.899542 4822 scope.go:117] "RemoveContainer" containerID="83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1" Nov 24 14:51:17 crc kubenswrapper[4822]: E1124 14:51:17.900011 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1\": container with ID starting with 83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1 not found: ID does not exist" containerID="83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.900034 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1"} err="failed to get container status \"83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1\": rpc error: code = NotFound desc = could not find container \"83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1\": container with ID starting with 83ceeecc8918f9ce35e0da0dbfa2fdc12e54aec35e28d7c4ea1e258b36a796b1 not found: ID does not exist" Nov 24 14:51:17 crc kubenswrapper[4822]: I1124 14:51:17.904065 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cbd4487cc-66fhk"] Nov 24 14:51:18 crc kubenswrapper[4822]: I1124 14:51:18.144440 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6b8fb4f557-bkwxl"] Nov 24 14:51:18 crc kubenswrapper[4822]: I1124 14:51:18.868046 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" event={"ID":"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf","Type":"ContainerStarted","Data":"08c8553ab373151bdc283ceca2d182291cf01a9d834bc3e32b4b77ee1e76d111"} Nov 24 14:51:18 crc kubenswrapper[4822]: I1124 14:51:18.868452 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" event={"ID":"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf","Type":"ContainerStarted","Data":"50a7e8826d7ed70d91425bb5ede087e107fef3c5cc0da3a7b707d13d5a2f3f51"} Nov 24 14:51:18 crc kubenswrapper[4822]: I1124 14:51:18.868489 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:51:18 crc kubenswrapper[4822]: I1124 14:51:18.868509 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" event={"ID":"753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf","Type":"ContainerStarted","Data":"9ad9b34d86c8f99f384fea5bae57f223e7e6e0b83ad2c2d797e9d29976e74476"} Nov 24 14:51:18 crc kubenswrapper[4822]: I1124 14:51:18.868529 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:51:18 crc kubenswrapper[4822]: I1124 14:51:18.903958 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" podStartSLOduration=739.903923841 podStartE2EDuration="12m19.903923841s" podCreationTimestamp="2025-11-24 14:38:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:51:18.898893871 +0000 UTC m=+1916.015534378" watchObservedRunningTime="2025-11-24 14:51:18.903923841 +0000 UTC m=+1916.020564358" Nov 24 14:51:19 crc kubenswrapper[4822]: I1124 14:51:19.718142 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02fe13aa-aece-4235-b424-e244d720b40f" path="/var/lib/kubelet/pods/02fe13aa-aece-4235-b424-e244d720b40f/volumes" Nov 24 14:51:27 crc kubenswrapper[4822]: I1124 14:51:27.612813 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:51:27 crc kubenswrapper[4822]: I1124 14:51:27.615830 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6b8fb4f557-bkwxl" Nov 24 14:51:28 crc kubenswrapper[4822]: I1124 14:51:28.850807 4822 scope.go:117] "RemoveContainer" containerID="b56c541e746656bb8ec65b3233905aaa6c9db78ea61d3d126e1ef7b3110c3306" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.497751 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-kks22"] Nov 24 14:51:38 crc kubenswrapper[4822]: E1124 14:51:38.498995 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fe13aa-aece-4235-b424-e244d720b40f" containerName="dnsmasq-dns" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.499018 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fe13aa-aece-4235-b424-e244d720b40f" containerName="dnsmasq-dns" Nov 24 14:51:38 crc kubenswrapper[4822]: E1124 14:51:38.499056 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fe13aa-aece-4235-b424-e244d720b40f" containerName="init" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.499069 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fe13aa-aece-4235-b424-e244d720b40f" containerName="init" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.499472 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="02fe13aa-aece-4235-b424-e244d720b40f" containerName="dnsmasq-dns" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.500686 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.503456 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.506079 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-kks22"] Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.658161 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-certs\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.658239 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-scripts\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.658327 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-combined-ca-bundle\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.658355 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-config-data\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.658471 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8jzn\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-kube-api-access-m8jzn\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.761279 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-combined-ca-bundle\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.761355 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-config-data\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.761402 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8jzn\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-kube-api-access-m8jzn\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.761582 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-certs\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.761640 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-scripts\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.770111 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-config-data\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.771018 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-scripts\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.775905 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-certs\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.783816 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8jzn\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-kube-api-access-m8jzn\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.784509 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-combined-ca-bundle\") pod \"cloudkitty-db-sync-kks22\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:38 crc kubenswrapper[4822]: I1124 14:51:38.818996 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:39 crc kubenswrapper[4822]: I1124 14:51:39.427715 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-kks22"] Nov 24 14:51:39 crc kubenswrapper[4822]: I1124 14:51:39.436262 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:51:40 crc kubenswrapper[4822]: I1124 14:51:40.145322 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-kks22" event={"ID":"f64cdb18-1ccf-46b4-bea2-9df357833a51","Type":"ContainerStarted","Data":"03c9e064b8818806a2256646070656f3060141f999369cc7d12aa82bd1574aac"} Nov 24 14:51:40 crc kubenswrapper[4822]: I1124 14:51:40.145938 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-kks22" event={"ID":"f64cdb18-1ccf-46b4-bea2-9df357833a51","Type":"ContainerStarted","Data":"cf4c873b0cae6f3b673591603e79c0108a11a656b00c1724d642cf6db13a2176"} Nov 24 14:51:40 crc kubenswrapper[4822]: I1124 14:51:40.167828 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-kks22" podStartSLOduration=1.9778364819999998 podStartE2EDuration="2.167810852s" podCreationTimestamp="2025-11-24 14:51:38 +0000 UTC" firstStartedPulling="2025-11-24 14:51:39.436057314 +0000 UTC m=+1936.552697791" lastFinishedPulling="2025-11-24 14:51:39.626031694 +0000 UTC m=+1936.742672161" observedRunningTime="2025-11-24 14:51:40.166605465 +0000 UTC m=+1937.283245982" watchObservedRunningTime="2025-11-24 14:51:40.167810852 +0000 UTC m=+1937.284451329" Nov 24 14:51:40 crc kubenswrapper[4822]: I1124 14:51:40.416501 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:51:40 crc kubenswrapper[4822]: I1124 14:51:40.670155 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:51:40 crc kubenswrapper[4822]: I1124 14:51:40.670530 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="proxy-httpd" containerID="cri-o://df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe" gracePeriod=30 Nov 24 14:51:40 crc kubenswrapper[4822]: I1124 14:51:40.670573 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="sg-core" containerID="cri-o://4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42" gracePeriod=30 Nov 24 14:51:40 crc kubenswrapper[4822]: I1124 14:51:40.670797 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="ceilometer-central-agent" containerID="cri-o://c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a" gracePeriod=30 Nov 24 14:51:40 crc kubenswrapper[4822]: I1124 14:51:40.670829 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="ceilometer-notification-agent" containerID="cri-o://325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301" gracePeriod=30 Nov 24 14:51:41 crc kubenswrapper[4822]: I1124 14:51:41.155812 4822 generic.go:334] "Generic (PLEG): container finished" podID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerID="df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe" exitCode=0 Nov 24 14:51:41 crc kubenswrapper[4822]: I1124 14:51:41.155850 4822 generic.go:334] "Generic (PLEG): container finished" podID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerID="4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42" exitCode=2 Nov 24 14:51:41 crc kubenswrapper[4822]: I1124 14:51:41.155859 4822 generic.go:334] "Generic (PLEG): container finished" podID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerID="c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a" exitCode=0 Nov 24 14:51:41 crc kubenswrapper[4822]: I1124 14:51:41.155888 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerDied","Data":"df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe"} Nov 24 14:51:41 crc kubenswrapper[4822]: I1124 14:51:41.155933 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerDied","Data":"4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42"} Nov 24 14:51:41 crc kubenswrapper[4822]: I1124 14:51:41.155944 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerDied","Data":"c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a"} Nov 24 14:51:41 crc kubenswrapper[4822]: I1124 14:51:41.318446 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.149875 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.173896 4822 generic.go:334] "Generic (PLEG): container finished" podID="f64cdb18-1ccf-46b4-bea2-9df357833a51" containerID="03c9e064b8818806a2256646070656f3060141f999369cc7d12aa82bd1574aac" exitCode=0 Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.173991 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-kks22" event={"ID":"f64cdb18-1ccf-46b4-bea2-9df357833a51","Type":"ContainerDied","Data":"03c9e064b8818806a2256646070656f3060141f999369cc7d12aa82bd1574aac"} Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.177093 4822 generic.go:334] "Generic (PLEG): container finished" podID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerID="325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301" exitCode=0 Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.177137 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerDied","Data":"325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301"} Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.177168 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35","Type":"ContainerDied","Data":"bce2d94199a0959f5cbe78669d333d7137dfacc85c99c152f6c2dc4b80d60f27"} Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.177178 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.177186 4822 scope.go:117] "RemoveContainer" containerID="df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.209522 4822 scope.go:117] "RemoveContainer" containerID="4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.245408 4822 scope.go:117] "RemoveContainer" containerID="325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.252104 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-sg-core-conf-yaml\") pod \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.252167 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-combined-ca-bundle\") pod \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.252223 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-ceilometer-tls-certs\") pod \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.252284 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-config-data\") pod \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.252336 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-scripts\") pod \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.252394 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-run-httpd\") pod \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.252460 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8fjd\" (UniqueName: \"kubernetes.io/projected/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-kube-api-access-l8fjd\") pod \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.253182 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-log-httpd\") pod \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\" (UID: \"33cb9a74-f7ff-485d-b820-2ef0f7eb4a35\") " Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.253376 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" (UID: "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.253802 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" (UID: "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.253843 4822 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.274873 4822 scope.go:117] "RemoveContainer" containerID="c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.300399 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-scripts" (OuterVolumeSpecName: "scripts") pod "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" (UID: "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.302050 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-kube-api-access-l8fjd" (OuterVolumeSpecName: "kube-api-access-l8fjd") pod "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" (UID: "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35"). InnerVolumeSpecName "kube-api-access-l8fjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.309868 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" (UID: "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.341333 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" (UID: "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.358263 4822 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.358293 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.358304 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8fjd\" (UniqueName: \"kubernetes.io/projected/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-kube-api-access-l8fjd\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.358315 4822 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.358323 4822 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.377376 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" (UID: "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.382560 4822 scope.go:117] "RemoveContainer" containerID="df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe" Nov 24 14:51:43 crc kubenswrapper[4822]: E1124 14:51:43.384680 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe\": container with ID starting with df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe not found: ID does not exist" containerID="df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.384724 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe"} err="failed to get container status \"df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe\": rpc error: code = NotFound desc = could not find container \"df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe\": container with ID starting with df7ec16b33edda048254b191d1a2efd326f8e216d5c0383302ed0718213e4bbe not found: ID does not exist" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.384751 4822 scope.go:117] "RemoveContainer" containerID="4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42" Nov 24 14:51:43 crc kubenswrapper[4822]: E1124 14:51:43.385073 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42\": container with ID starting with 4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42 not found: ID does not exist" containerID="4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.385114 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42"} err="failed to get container status \"4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42\": rpc error: code = NotFound desc = could not find container \"4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42\": container with ID starting with 4e488c285420a014deee45cb888444faf25c2a69dcf0ccd329cc2ed09bb3fb42 not found: ID does not exist" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.385133 4822 scope.go:117] "RemoveContainer" containerID="325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301" Nov 24 14:51:43 crc kubenswrapper[4822]: E1124 14:51:43.385654 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301\": container with ID starting with 325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301 not found: ID does not exist" containerID="325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.385684 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301"} err="failed to get container status \"325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301\": rpc error: code = NotFound desc = could not find container \"325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301\": container with ID starting with 325ca11e683adba6e3c0b8c61c12564e9d12d25c92e3bca316b8b721edc75301 not found: ID does not exist" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.385702 4822 scope.go:117] "RemoveContainer" containerID="c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a" Nov 24 14:51:43 crc kubenswrapper[4822]: E1124 14:51:43.386422 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a\": container with ID starting with c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a not found: ID does not exist" containerID="c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.386460 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a"} err="failed to get container status \"c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a\": rpc error: code = NotFound desc = could not find container \"c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a\": container with ID starting with c28a0c5293dd1f1970c19ca3697c6ae27f108e72190288186f3b2b9b5cd25a1a not found: ID does not exist" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.436280 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-config-data" (OuterVolumeSpecName: "config-data") pod "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" (UID: "33cb9a74-f7ff-485d-b820-2ef0f7eb4a35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.460233 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.460262 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.514003 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.521576 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.541326 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:51:43 crc kubenswrapper[4822]: E1124 14:51:43.541699 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="ceilometer-notification-agent" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.541716 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="ceilometer-notification-agent" Nov 24 14:51:43 crc kubenswrapper[4822]: E1124 14:51:43.541741 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="sg-core" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.541747 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="sg-core" Nov 24 14:51:43 crc kubenswrapper[4822]: E1124 14:51:43.541766 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="proxy-httpd" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.541772 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="proxy-httpd" Nov 24 14:51:43 crc kubenswrapper[4822]: E1124 14:51:43.541794 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="ceilometer-central-agent" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.541801 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="ceilometer-central-agent" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.541969 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="proxy-httpd" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.541991 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="ceilometer-notification-agent" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.542002 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="ceilometer-central-agent" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.542013 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" containerName="sg-core" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.543834 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.545915 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.546108 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.547224 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.569451 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.663190 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.663266 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7891e240-53f5-434a-9ee1-8e3362ac4686-run-httpd\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.663577 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-config-data\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.663770 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-scripts\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.663809 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.663884 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7891e240-53f5-434a-9ee1-8e3362ac4686-log-httpd\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.663906 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq9fn\" (UniqueName: \"kubernetes.io/projected/7891e240-53f5-434a-9ee1-8e3362ac4686-kube-api-access-bq9fn\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.663928 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.719367 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33cb9a74-f7ff-485d-b820-2ef0f7eb4a35" path="/var/lib/kubelet/pods/33cb9a74-f7ff-485d-b820-2ef0f7eb4a35/volumes" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.766343 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.766394 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7891e240-53f5-434a-9ee1-8e3362ac4686-run-httpd\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.766476 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-config-data\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.766517 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-scripts\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.766533 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.766563 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7891e240-53f5-434a-9ee1-8e3362ac4686-log-httpd\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.766580 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq9fn\" (UniqueName: \"kubernetes.io/projected/7891e240-53f5-434a-9ee1-8e3362ac4686-kube-api-access-bq9fn\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.766602 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.767222 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7891e240-53f5-434a-9ee1-8e3362ac4686-run-httpd\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.767312 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7891e240-53f5-434a-9ee1-8e3362ac4686-log-httpd\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.770588 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.771126 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-config-data\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.772833 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.774904 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-scripts\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.775461 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7891e240-53f5-434a-9ee1-8e3362ac4686-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.797884 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq9fn\" (UniqueName: \"kubernetes.io/projected/7891e240-53f5-434a-9ee1-8e3362ac4686-kube-api-access-bq9fn\") pod \"ceilometer-0\" (UID: \"7891e240-53f5-434a-9ee1-8e3362ac4686\") " pod="openstack/ceilometer-0" Nov 24 14:51:43 crc kubenswrapper[4822]: I1124 14:51:43.858137 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.389254 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.490379 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.585575 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8jzn\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-kube-api-access-m8jzn\") pod \"f64cdb18-1ccf-46b4-bea2-9df357833a51\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.585651 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-config-data\") pod \"f64cdb18-1ccf-46b4-bea2-9df357833a51\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.585884 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-scripts\") pod \"f64cdb18-1ccf-46b4-bea2-9df357833a51\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.585941 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-certs\") pod \"f64cdb18-1ccf-46b4-bea2-9df357833a51\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.585973 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-combined-ca-bundle\") pod \"f64cdb18-1ccf-46b4-bea2-9df357833a51\" (UID: \"f64cdb18-1ccf-46b4-bea2-9df357833a51\") " Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.591495 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-scripts" (OuterVolumeSpecName: "scripts") pod "f64cdb18-1ccf-46b4-bea2-9df357833a51" (UID: "f64cdb18-1ccf-46b4-bea2-9df357833a51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.592522 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-certs" (OuterVolumeSpecName: "certs") pod "f64cdb18-1ccf-46b4-bea2-9df357833a51" (UID: "f64cdb18-1ccf-46b4-bea2-9df357833a51"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.592569 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-kube-api-access-m8jzn" (OuterVolumeSpecName: "kube-api-access-m8jzn") pod "f64cdb18-1ccf-46b4-bea2-9df357833a51" (UID: "f64cdb18-1ccf-46b4-bea2-9df357833a51"). InnerVolumeSpecName "kube-api-access-m8jzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.617797 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-config-data" (OuterVolumeSpecName: "config-data") pod "f64cdb18-1ccf-46b4-bea2-9df357833a51" (UID: "f64cdb18-1ccf-46b4-bea2-9df357833a51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.621187 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f64cdb18-1ccf-46b4-bea2-9df357833a51" (UID: "f64cdb18-1ccf-46b4-bea2-9df357833a51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.688527 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8jzn\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-kube-api-access-m8jzn\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.688577 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.688587 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.688595 4822 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f64cdb18-1ccf-46b4-bea2-9df357833a51-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.688603 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64cdb18-1ccf-46b4-bea2-9df357833a51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:44 crc kubenswrapper[4822]: I1124 14:51:44.779648 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" containerName="rabbitmq" containerID="cri-o://dec06a7f92f37aaabc29aea08f1da7e2195f934ec84abde199e37f03bf023343" gracePeriod=604796 Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.231131 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7891e240-53f5-434a-9ee1-8e3362ac4686","Type":"ContainerStarted","Data":"d7454f897deedee7bae46b9cd80c72ab001f1eda61467b8542e999da3336af85"} Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.247170 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-kks22" event={"ID":"f64cdb18-1ccf-46b4-bea2-9df357833a51","Type":"ContainerDied","Data":"cf4c873b0cae6f3b673591603e79c0108a11a656b00c1724d642cf6db13a2176"} Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.247225 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf4c873b0cae6f3b673591603e79c0108a11a656b00c1724d642cf6db13a2176" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.247293 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-kks22" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.345195 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-v5pw2"] Nov 24 14:51:45 crc kubenswrapper[4822]: E1124 14:51:45.345612 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f64cdb18-1ccf-46b4-bea2-9df357833a51" containerName="cloudkitty-db-sync" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.345630 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f64cdb18-1ccf-46b4-bea2-9df357833a51" containerName="cloudkitty-db-sync" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.345838 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="f64cdb18-1ccf-46b4-bea2-9df357833a51" containerName="cloudkitty-db-sync" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.346620 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.349642 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.367616 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-v5pw2"] Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.529958 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-combined-ca-bundle\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.530015 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvfm4\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-kube-api-access-bvfm4\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.530098 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-scripts\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.530158 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-certs\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.530183 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-config-data\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.632086 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-combined-ca-bundle\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.632142 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvfm4\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-kube-api-access-bvfm4\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.632185 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-scripts\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.632267 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-certs\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.632289 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-config-data\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.637966 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-scripts\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.638146 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-certs\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.642381 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-config-data\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.653843 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-combined-ca-bundle\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.653976 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvfm4\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-kube-api-access-bvfm4\") pod \"cloudkitty-storageinit-v5pw2\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:45 crc kubenswrapper[4822]: I1124 14:51:45.684804 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:46 crc kubenswrapper[4822]: I1124 14:51:46.211831 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-v5pw2"] Nov 24 14:51:46 crc kubenswrapper[4822]: I1124 14:51:46.264540 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="a52d37ed-a00f-4983-88ee-023bd282038e" containerName="rabbitmq" containerID="cri-o://48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963" gracePeriod=604796 Nov 24 14:51:47 crc kubenswrapper[4822]: W1124 14:51:47.987827 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ed26986_cc75_4e38_b2ab_53b839cbb8d6.slice/crio-600018b4d72294b7d9aee0cbdbd0bed7bc456669e47170e99fbbea137761bb7f WatchSource:0}: Error finding container 600018b4d72294b7d9aee0cbdbd0bed7bc456669e47170e99fbbea137761bb7f: Status 404 returned error can't find the container with id 600018b4d72294b7d9aee0cbdbd0bed7bc456669e47170e99fbbea137761bb7f Nov 24 14:51:48 crc kubenswrapper[4822]: I1124 14:51:48.282887 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-v5pw2" event={"ID":"9ed26986-cc75-4e38-b2ab-53b839cbb8d6","Type":"ContainerStarted","Data":"600018b4d72294b7d9aee0cbdbd0bed7bc456669e47170e99fbbea137761bb7f"} Nov 24 14:51:49 crc kubenswrapper[4822]: I1124 14:51:49.315989 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-v5pw2" event={"ID":"9ed26986-cc75-4e38-b2ab-53b839cbb8d6","Type":"ContainerStarted","Data":"6fe9ae59092417bc7ce7646da5ccc38209363d4e24f04c4de0f4349ceac307fc"} Nov 24 14:51:49 crc kubenswrapper[4822]: I1124 14:51:49.319383 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7891e240-53f5-434a-9ee1-8e3362ac4686","Type":"ContainerStarted","Data":"94e5f35602abd60797134a2882529d12306b1abffb988f909064939b9d3dd80f"} Nov 24 14:51:49 crc kubenswrapper[4822]: I1124 14:51:49.319426 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7891e240-53f5-434a-9ee1-8e3362ac4686","Type":"ContainerStarted","Data":"27401dbcbb0510da8036777940d6c00efc55a8d122e4049db19164d05928b427"} Nov 24 14:51:49 crc kubenswrapper[4822]: I1124 14:51:49.344340 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-v5pw2" podStartSLOduration=4.344315441 podStartE2EDuration="4.344315441s" podCreationTimestamp="2025-11-24 14:51:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:51:49.337690402 +0000 UTC m=+1946.454330919" watchObservedRunningTime="2025-11-24 14:51:49.344315441 +0000 UTC m=+1946.460955948" Nov 24 14:51:50 crc kubenswrapper[4822]: I1124 14:51:50.337773 4822 generic.go:334] "Generic (PLEG): container finished" podID="9ed26986-cc75-4e38-b2ab-53b839cbb8d6" containerID="6fe9ae59092417bc7ce7646da5ccc38209363d4e24f04c4de0f4349ceac307fc" exitCode=0 Nov 24 14:51:50 crc kubenswrapper[4822]: I1124 14:51:50.338180 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-v5pw2" event={"ID":"9ed26986-cc75-4e38-b2ab-53b839cbb8d6","Type":"ContainerDied","Data":"6fe9ae59092417bc7ce7646da5ccc38209363d4e24f04c4de0f4349ceac307fc"} Nov 24 14:51:50 crc kubenswrapper[4822]: I1124 14:51:50.341759 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7891e240-53f5-434a-9ee1-8e3362ac4686","Type":"ContainerStarted","Data":"dbaddfe59ba63a5cf743095148b76060512c1b26c73898fd2b4f6d23a829ecf7"} Nov 24 14:51:50 crc kubenswrapper[4822]: I1124 14:51:50.725944 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.110:5671: connect: connection refused" Nov 24 14:51:50 crc kubenswrapper[4822]: I1124 14:51:50.779313 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="a52d37ed-a00f-4983-88ee-023bd282038e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.111:5671: connect: connection refused" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.353678 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7891e240-53f5-434a-9ee1-8e3362ac4686","Type":"ContainerStarted","Data":"232ca36ec6968035a695b47079ef413f7ba658a704b19471f58e25185b391a80"} Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.354296 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.356753 4822 generic.go:334] "Generic (PLEG): container finished" podID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" containerID="dec06a7f92f37aaabc29aea08f1da7e2195f934ec84abde199e37f03bf023343" exitCode=0 Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.357856 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37","Type":"ContainerDied","Data":"dec06a7f92f37aaabc29aea08f1da7e2195f934ec84abde199e37f03bf023343"} Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.382312 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.776103906 podStartE2EDuration="8.382297583s" podCreationTimestamp="2025-11-24 14:51:43 +0000 UTC" firstStartedPulling="2025-11-24 14:51:44.398968026 +0000 UTC m=+1941.515608503" lastFinishedPulling="2025-11-24 14:51:51.005161673 +0000 UTC m=+1948.121802180" observedRunningTime="2025-11-24 14:51:51.373047661 +0000 UTC m=+1948.489688138" watchObservedRunningTime="2025-11-24 14:51:51.382297583 +0000 UTC m=+1948.498938060" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.528713 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.718809 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.718856 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-erlang-cookie-secret\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.718896 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-config-data\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.718988 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-server-conf\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.719053 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-plugins\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.719068 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-confd\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.719105 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-erlang-cookie\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.719150 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlljz\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-kube-api-access-nlljz\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.719180 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-pod-info\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.719239 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-plugins-conf\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.719281 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-tls\") pod \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\" (UID: \"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.728835 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.732464 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.743381 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.746773 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.747013 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.754369 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-kube-api-access-nlljz" (OuterVolumeSpecName: "kube-api-access-nlljz") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "kube-api-access-nlljz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.768890 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-pod-info" (OuterVolumeSpecName: "pod-info") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.800734 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-config-data" (OuterVolumeSpecName: "config-data") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.818420 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687" (OuterVolumeSpecName: "persistence") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "pvc-f386849d-6cef-4916-a1ab-849b5b93b687". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.822171 4822 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.822225 4822 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") on node \"crc\" " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.822238 4822 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.822248 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.822257 4822 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.822266 4822 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.822275 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlljz\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-kube-api-access-nlljz\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.822283 4822 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.822291 4822 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.862847 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-server-conf" (OuterVolumeSpecName: "server-conf") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.884535 4822 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.884703 4822 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f386849d-6cef-4916-a1ab-849b5b93b687" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687") on node "crc" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.898267 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.924265 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" (UID: "f3571ac5-4881-4dfc-aaa2-60d7c5c39a37"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.924329 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-config-data\") pod \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.924639 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-certs\") pod \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.924778 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-scripts\") pod \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.924884 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvfm4\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-kube-api-access-bvfm4\") pod \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.925020 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-combined-ca-bundle\") pod \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\" (UID: \"9ed26986-cc75-4e38-b2ab-53b839cbb8d6\") " Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.925759 4822 reconciler_common.go:293] "Volume detached for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.926197 4822 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.926321 4822 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.928598 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-kube-api-access-bvfm4" (OuterVolumeSpecName: "kube-api-access-bvfm4") pod "9ed26986-cc75-4e38-b2ab-53b839cbb8d6" (UID: "9ed26986-cc75-4e38-b2ab-53b839cbb8d6"). InnerVolumeSpecName "kube-api-access-bvfm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.934270 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-certs" (OuterVolumeSpecName: "certs") pod "9ed26986-cc75-4e38-b2ab-53b839cbb8d6" (UID: "9ed26986-cc75-4e38-b2ab-53b839cbb8d6"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.934386 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-scripts" (OuterVolumeSpecName: "scripts") pod "9ed26986-cc75-4e38-b2ab-53b839cbb8d6" (UID: "9ed26986-cc75-4e38-b2ab-53b839cbb8d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.961324 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ed26986-cc75-4e38-b2ab-53b839cbb8d6" (UID: "9ed26986-cc75-4e38-b2ab-53b839cbb8d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:51 crc kubenswrapper[4822]: I1124 14:51:51.978561 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-config-data" (OuterVolumeSpecName: "config-data") pod "9ed26986-cc75-4e38-b2ab-53b839cbb8d6" (UID: "9ed26986-cc75-4e38-b2ab-53b839cbb8d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.031664 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvfm4\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-kube-api-access-bvfm4\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.031696 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.031705 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.031715 4822 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.031723 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed26986-cc75-4e38-b2ab-53b839cbb8d6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.367743 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-v5pw2" event={"ID":"9ed26986-cc75-4e38-b2ab-53b839cbb8d6","Type":"ContainerDied","Data":"600018b4d72294b7d9aee0cbdbd0bed7bc456669e47170e99fbbea137761bb7f"} Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.367791 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="600018b4d72294b7d9aee0cbdbd0bed7bc456669e47170e99fbbea137761bb7f" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.367855 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-v5pw2" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.379503 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.379529 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3571ac5-4881-4dfc-aaa2-60d7c5c39a37","Type":"ContainerDied","Data":"5af2c0566f682b6183a41fec80b61b517b204d27891cd6a77a63ee67e7ff6060"} Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.379581 4822 scope.go:117] "RemoveContainer" containerID="dec06a7f92f37aaabc29aea08f1da7e2195f934ec84abde199e37f03bf023343" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.481258 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.481792 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="74794267-7a2e-4dc4-bb6b-35b99814e3ee" containerName="cloudkitty-proc" containerID="cri-o://48c41eaf640975344f69e7f1b14f82dce5425912c516ece0728ca67e1715543d" gracePeriod=30 Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.496005 4822 scope.go:117] "RemoveContainer" containerID="a11d1a87ad59947fd0c7dafe5c9ace40accaa87c88506ddfd156b36fa2cfb257" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.510260 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.521683 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.537498 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:51:52 crc kubenswrapper[4822]: E1124 14:51:52.537921 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" containerName="rabbitmq" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.537940 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" containerName="rabbitmq" Nov 24 14:51:52 crc kubenswrapper[4822]: E1124 14:51:52.537983 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed26986-cc75-4e38-b2ab-53b839cbb8d6" containerName="cloudkitty-storageinit" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.537990 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed26986-cc75-4e38-b2ab-53b839cbb8d6" containerName="cloudkitty-storageinit" Nov 24 14:51:52 crc kubenswrapper[4822]: E1124 14:51:52.538011 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" containerName="setup-container" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.538018 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" containerName="setup-container" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.538185 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed26986-cc75-4e38-b2ab-53b839cbb8d6" containerName="cloudkitty-storageinit" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.538223 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" containerName="rabbitmq" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.539473 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.542481 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.546032 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vc8w9" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.546289 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.546421 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.546531 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.546665 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.550462 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.564243 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.574483 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.574758 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerName="cloudkitty-api-log" containerID="cri-o://0ed5ee75ecb4ed8f0fbfe3dff427d71463d18ecc6ae018bce84cf9fcbe2de39b" gracePeriod=30 Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.575138 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerName="cloudkitty-api" containerID="cri-o://a8fe55b14cbb0e20e3997f9943574c2356792e3425cd3798ec3cc8bfa3a82050" gracePeriod=30 Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.651861 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.651907 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af242c06-68a2-436b-84d9-c2d020f15903-pod-info\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.651952 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af242c06-68a2-436b-84d9-c2d020f15903-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.651976 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.652017 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af242c06-68a2-436b-84d9-c2d020f15903-config-data\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.652059 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.652075 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af242c06-68a2-436b-84d9-c2d020f15903-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.652092 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af242c06-68a2-436b-84d9-c2d020f15903-server-conf\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.652153 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.652171 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.652194 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8244l\" (UniqueName: \"kubernetes.io/projected/af242c06-68a2-436b-84d9-c2d020f15903-kube-api-access-8244l\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.753891 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.753947 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.753981 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8244l\" (UniqueName: \"kubernetes.io/projected/af242c06-68a2-436b-84d9-c2d020f15903-kube-api-access-8244l\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.754073 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.754095 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af242c06-68a2-436b-84d9-c2d020f15903-pod-info\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.754135 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af242c06-68a2-436b-84d9-c2d020f15903-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.754159 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.754217 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af242c06-68a2-436b-84d9-c2d020f15903-config-data\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.754253 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.754280 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af242c06-68a2-436b-84d9-c2d020f15903-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.754297 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af242c06-68a2-436b-84d9-c2d020f15903-server-conf\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.755588 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.755710 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af242c06-68a2-436b-84d9-c2d020f15903-server-conf\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.756768 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af242c06-68a2-436b-84d9-c2d020f15903-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.758551 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.761163 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af242c06-68a2-436b-84d9-c2d020f15903-config-data\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.761881 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.761907 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0ec351afa2b1df6aac252f23371d09947c6973707300212e509127eecca4cbe8/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.762917 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.762965 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af242c06-68a2-436b-84d9-c2d020f15903-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.765415 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af242c06-68a2-436b-84d9-c2d020f15903-pod-info\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.766104 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af242c06-68a2-436b-84d9-c2d020f15903-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.771899 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8244l\" (UniqueName: \"kubernetes.io/projected/af242c06-68a2-436b-84d9-c2d020f15903-kube-api-access-8244l\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.863170 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f386849d-6cef-4916-a1ab-849b5b93b687\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f386849d-6cef-4916-a1ab-849b5b93b687\") pod \"rabbitmq-server-0\" (UID: \"af242c06-68a2-436b-84d9-c2d020f15903\") " pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.873568 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.921340 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.960038 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a52d37ed-a00f-4983-88ee-023bd282038e-pod-info\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.960268 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a52d37ed-a00f-4983-88ee-023bd282038e-erlang-cookie-secret\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.960313 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-server-conf\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.960330 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-config-data\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.960367 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw4w4\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-kube-api-access-qw4w4\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.960418 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-confd\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.960441 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-plugins-conf\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.960499 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-plugins\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.960551 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-erlang-cookie\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.961494 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.961547 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-tls\") pod \"a52d37ed-a00f-4983-88ee-023bd282038e\" (UID: \"a52d37ed-a00f-4983-88ee-023bd282038e\") " Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.963288 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.963959 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.965007 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a52d37ed-a00f-4983-88ee-023bd282038e-pod-info" (OuterVolumeSpecName: "pod-info") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.965776 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.968603 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-kube-api-access-qw4w4" (OuterVolumeSpecName: "kube-api-access-qw4w4") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "kube-api-access-qw4w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.969348 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.979818 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a52d37ed-a00f-4983-88ee-023bd282038e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:52 crc kubenswrapper[4822]: I1124 14:51:52.992609 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594" (OuterVolumeSpecName: "persistence") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "pvc-4a284a73-8a54-415c-b953-7cef6a044594". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.024911 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-config-data" (OuterVolumeSpecName: "config-data") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.065491 4822 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.065522 4822 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a52d37ed-a00f-4983-88ee-023bd282038e-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.065531 4822 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a52d37ed-a00f-4983-88ee-023bd282038e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.065542 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.065552 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw4w4\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-kube-api-access-qw4w4\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.065561 4822 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.065569 4822 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.065578 4822 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.065608 4822 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") on node \"crc\" " Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.099821 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-server-conf" (OuterVolumeSpecName: "server-conf") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.126955 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a52d37ed-a00f-4983-88ee-023bd282038e" (UID: "a52d37ed-a00f-4983-88ee-023bd282038e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.168576 4822 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a52d37ed-a00f-4983-88ee-023bd282038e-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.168606 4822 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a52d37ed-a00f-4983-88ee-023bd282038e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.187657 4822 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.187833 4822 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-4a284a73-8a54-415c-b953-7cef6a044594" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594") on node "crc" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.270828 4822 reconciler_common.go:293] "Volume detached for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.395010 4822 generic.go:334] "Generic (PLEG): container finished" podID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerID="0ed5ee75ecb4ed8f0fbfe3dff427d71463d18ecc6ae018bce84cf9fcbe2de39b" exitCode=143 Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.395072 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"51daf6fb-63aa-4ab7-8e3b-69e861e67e43","Type":"ContainerDied","Data":"0ed5ee75ecb4ed8f0fbfe3dff427d71463d18ecc6ae018bce84cf9fcbe2de39b"} Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.397221 4822 generic.go:334] "Generic (PLEG): container finished" podID="74794267-7a2e-4dc4-bb6b-35b99814e3ee" containerID="48c41eaf640975344f69e7f1b14f82dce5425912c516ece0728ca67e1715543d" exitCode=0 Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.397239 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"74794267-7a2e-4dc4-bb6b-35b99814e3ee","Type":"ContainerDied","Data":"48c41eaf640975344f69e7f1b14f82dce5425912c516ece0728ca67e1715543d"} Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.398997 4822 generic.go:334] "Generic (PLEG): container finished" podID="a52d37ed-a00f-4983-88ee-023bd282038e" containerID="48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963" exitCode=0 Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.399027 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a52d37ed-a00f-4983-88ee-023bd282038e","Type":"ContainerDied","Data":"48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963"} Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.399051 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a52d37ed-a00f-4983-88ee-023bd282038e","Type":"ContainerDied","Data":"3ef59d208b26c421729e202a5dc4b21edcaa051120d52d21eedcab90fb1fafff"} Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.399067 4822 scope.go:117] "RemoveContainer" containerID="48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.399186 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.434635 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.443337 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.445179 4822 scope.go:117] "RemoveContainer" containerID="276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.474311 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:51:53 crc kubenswrapper[4822]: E1124 14:51:53.474727 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a52d37ed-a00f-4983-88ee-023bd282038e" containerName="setup-container" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.474741 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a52d37ed-a00f-4983-88ee-023bd282038e" containerName="setup-container" Nov 24 14:51:53 crc kubenswrapper[4822]: E1124 14:51:53.474772 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a52d37ed-a00f-4983-88ee-023bd282038e" containerName="rabbitmq" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.474779 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a52d37ed-a00f-4983-88ee-023bd282038e" containerName="rabbitmq" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.474973 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a52d37ed-a00f-4983-88ee-023bd282038e" containerName="rabbitmq" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.476458 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.479455 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.481735 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.481883 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.481993 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-bzdzc" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.482118 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.482238 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.483869 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.496908 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.531424 4822 scope.go:117] "RemoveContainer" containerID="48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963" Nov 24 14:51:53 crc kubenswrapper[4822]: E1124 14:51:53.535339 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963\": container with ID starting with 48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963 not found: ID does not exist" containerID="48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.535385 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963"} err="failed to get container status \"48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963\": rpc error: code = NotFound desc = could not find container \"48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963\": container with ID starting with 48a069250dabfa0c780dd2e82c327f16cfb9afc773a01aeacc9cb897924f0963 not found: ID does not exist" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.535418 4822 scope.go:117] "RemoveContainer" containerID="276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe" Nov 24 14:51:53 crc kubenswrapper[4822]: E1124 14:51:53.538449 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe\": container with ID starting with 276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe not found: ID does not exist" containerID="276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.538497 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe"} err="failed to get container status \"276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe\": rpc error: code = NotFound desc = could not find container \"276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe\": container with ID starting with 276c48b426adaba421a89f11de85b0ba826bfeb1d40a243cccea816b678592fe not found: ID does not exist" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.557255 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.575819 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c969e517-0251-4537-9999-53bf0cfa7cf0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.575875 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.575912 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.575938 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c969e517-0251-4537-9999-53bf0cfa7cf0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.575958 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c969e517-0251-4537-9999-53bf0cfa7cf0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.575981 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.576017 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.576064 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c969e517-0251-4537-9999-53bf0cfa7cf0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.576146 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.576172 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c969e517-0251-4537-9999-53bf0cfa7cf0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.576189 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmjw6\" (UniqueName: \"kubernetes.io/projected/c969e517-0251-4537-9999-53bf0cfa7cf0-kube-api-access-nmjw6\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.679830 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680143 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c969e517-0251-4537-9999-53bf0cfa7cf0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680175 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmjw6\" (UniqueName: \"kubernetes.io/projected/c969e517-0251-4537-9999-53bf0cfa7cf0-kube-api-access-nmjw6\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680286 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c969e517-0251-4537-9999-53bf0cfa7cf0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680322 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680361 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680385 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c969e517-0251-4537-9999-53bf0cfa7cf0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680403 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c969e517-0251-4537-9999-53bf0cfa7cf0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680427 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680466 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.680510 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c969e517-0251-4537-9999-53bf0cfa7cf0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.681156 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.681319 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c969e517-0251-4537-9999-53bf0cfa7cf0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.681847 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.682312 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c969e517-0251-4537-9999-53bf0cfa7cf0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.682462 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c969e517-0251-4537-9999-53bf0cfa7cf0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.686117 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c969e517-0251-4537-9999-53bf0cfa7cf0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.687407 4822 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.687438 4822 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c6993c052f1ba2e354c4a7c732feee88a39bd89c27ed6adb11f5ec90b4236450/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.688435 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.692765 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c969e517-0251-4537-9999-53bf0cfa7cf0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.693281 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c969e517-0251-4537-9999-53bf0cfa7cf0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.700041 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmjw6\" (UniqueName: \"kubernetes.io/projected/c969e517-0251-4537-9999-53bf0cfa7cf0-kube-api-access-nmjw6\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.716957 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a52d37ed-a00f-4983-88ee-023bd282038e" path="/var/lib/kubelet/pods/a52d37ed-a00f-4983-88ee-023bd282038e/volumes" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.718511 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3571ac5-4881-4dfc-aaa2-60d7c5c39a37" path="/var/lib/kubelet/pods/f3571ac5-4881-4dfc-aaa2-60d7c5c39a37/volumes" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.719316 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.731590 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4a284a73-8a54-415c-b953-7cef6a044594\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4a284a73-8a54-415c-b953-7cef6a044594\") pod \"rabbitmq-cell1-server-0\" (UID: \"c969e517-0251-4537-9999-53bf0cfa7cf0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.781528 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-combined-ca-bundle\") pod \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.781759 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkgw7\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-kube-api-access-tkgw7\") pod \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.781796 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-scripts\") pod \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.781853 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data\") pod \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.781883 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-certs\") pod \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.781907 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data-custom\") pod \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\" (UID: \"74794267-7a2e-4dc4-bb6b-35b99814e3ee\") " Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.785294 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-scripts" (OuterVolumeSpecName: "scripts") pod "74794267-7a2e-4dc4-bb6b-35b99814e3ee" (UID: "74794267-7a2e-4dc4-bb6b-35b99814e3ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.786243 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-kube-api-access-tkgw7" (OuterVolumeSpecName: "kube-api-access-tkgw7") pod "74794267-7a2e-4dc4-bb6b-35b99814e3ee" (UID: "74794267-7a2e-4dc4-bb6b-35b99814e3ee"). InnerVolumeSpecName "kube-api-access-tkgw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.786807 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-certs" (OuterVolumeSpecName: "certs") pod "74794267-7a2e-4dc4-bb6b-35b99814e3ee" (UID: "74794267-7a2e-4dc4-bb6b-35b99814e3ee"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.787594 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "74794267-7a2e-4dc4-bb6b-35b99814e3ee" (UID: "74794267-7a2e-4dc4-bb6b-35b99814e3ee"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.809837 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74794267-7a2e-4dc4-bb6b-35b99814e3ee" (UID: "74794267-7a2e-4dc4-bb6b-35b99814e3ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.812367 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data" (OuterVolumeSpecName: "config-data") pod "74794267-7a2e-4dc4-bb6b-35b99814e3ee" (UID: "74794267-7a2e-4dc4-bb6b-35b99814e3ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.856856 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.884619 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.884649 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkgw7\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-kube-api-access-tkgw7\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.884660 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.884669 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.884676 4822 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/74794267-7a2e-4dc4-bb6b-35b99814e3ee-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:53 crc kubenswrapper[4822]: I1124 14:51:53.884685 4822 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74794267-7a2e-4dc4-bb6b-35b99814e3ee-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.311231 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:51:54 crc kubenswrapper[4822]: W1124 14:51:54.323008 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc969e517_0251_4537_9999_53bf0cfa7cf0.slice/crio-0e618454ce3c18e915f6b1e4f1e61e361ec0acb661c11f585e58c8e58676690b WatchSource:0}: Error finding container 0e618454ce3c18e915f6b1e4f1e61e361ec0acb661c11f585e58c8e58676690b: Status 404 returned error can't find the container with id 0e618454ce3c18e915f6b1e4f1e61e361ec0acb661c11f585e58c8e58676690b Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.430555 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c969e517-0251-4537-9999-53bf0cfa7cf0","Type":"ContainerStarted","Data":"0e618454ce3c18e915f6b1e4f1e61e361ec0acb661c11f585e58c8e58676690b"} Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.436292 4822 generic.go:334] "Generic (PLEG): container finished" podID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerID="a8fe55b14cbb0e20e3997f9943574c2356792e3425cd3798ec3cc8bfa3a82050" exitCode=0 Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.436345 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"51daf6fb-63aa-4ab7-8e3b-69e861e67e43","Type":"ContainerDied","Data":"a8fe55b14cbb0e20e3997f9943574c2356792e3425cd3798ec3cc8bfa3a82050"} Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.439706 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"74794267-7a2e-4dc4-bb6b-35b99814e3ee","Type":"ContainerDied","Data":"ac2d57e7f707b8d3babbd443fd878be465036b914de6d34d52ff777ad9e9d536"} Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.439738 4822 scope.go:117] "RemoveContainer" containerID="48c41eaf640975344f69e7f1b14f82dce5425912c516ece0728ca67e1715543d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.439818 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.447590 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af242c06-68a2-436b-84d9-c2d020f15903","Type":"ContainerStarted","Data":"e1117ed2276d8b77515d44215c04ad8dbbdbe892e5b9133686981f2f7acfdda1"} Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.621850 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-584fd7dfd7-qdq77"] Nov 24 14:51:54 crc kubenswrapper[4822]: E1124 14:51:54.622502 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74794267-7a2e-4dc4-bb6b-35b99814e3ee" containerName="cloudkitty-proc" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.622516 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="74794267-7a2e-4dc4-bb6b-35b99814e3ee" containerName="cloudkitty-proc" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.624330 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="74794267-7a2e-4dc4-bb6b-35b99814e3ee" containerName="cloudkitty-proc" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.626495 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.634472 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-584fd7dfd7-qdq77"] Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.636463 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.750665 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-584fd7dfd7-qdq77"] Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.775235 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bgmw\" (UniqueName: \"kubernetes.io/projected/b79568a5-e8f5-479c-9978-d55609cb5f59-kube-api-access-9bgmw\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.775284 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-openstack-edpm-ipam\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.775351 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-svc\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.775371 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-swift-storage-0\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.775396 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-sb\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.775424 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-config\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.775447 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-nb\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.784433 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79c5f76cb7-gj65d"] Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.786158 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.795355 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79c5f76cb7-gj65d"] Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877359 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-swift-storage-0\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877403 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-sb\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877433 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-config\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877460 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-config\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877474 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-nb\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877531 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-openstack-edpm-ipam\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877550 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-ovsdbserver-nb\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877603 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-ovsdbserver-sb\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877637 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bgmw\" (UniqueName: \"kubernetes.io/projected/b79568a5-e8f5-479c-9978-d55609cb5f59-kube-api-access-9bgmw\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877661 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-openstack-edpm-ipam\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877677 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4kq9\" (UniqueName: \"kubernetes.io/projected/62305997-e4f9-4e0e-9e28-0dea4382bf40-kube-api-access-g4kq9\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877716 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-dns-svc\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877755 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-svc\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.877775 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-dns-swift-storage-0\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.878732 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-swift-storage-0\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.879253 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-sb\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.879800 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-config\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.880116 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-nb\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.880313 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-openstack-edpm-ipam\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.880747 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-svc\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: E1124 14:51:54.927278 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-9bgmw], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" podUID="b79568a5-e8f5-479c-9978-d55609cb5f59" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.944256 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.953033 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.959371 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.982048 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-ovsdbserver-sb\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.982139 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4kq9\" (UniqueName: \"kubernetes.io/projected/62305997-e4f9-4e0e-9e28-0dea4382bf40-kube-api-access-g4kq9\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.982198 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-dns-svc\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.982374 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-dns-swift-storage-0\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.982423 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-config\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.982517 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-openstack-edpm-ipam\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.982538 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-ovsdbserver-nb\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.983177 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-ovsdbserver-sb\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.983688 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-ovsdbserver-nb\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.983902 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-dns-swift-storage-0\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.986433 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-config\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.986517 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bgmw\" (UniqueName: \"kubernetes.io/projected/b79568a5-e8f5-479c-9978-d55609cb5f59-kube-api-access-9bgmw\") pod \"dnsmasq-dns-584fd7dfd7-qdq77\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.987398 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-openstack-edpm-ipam\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:54 crc kubenswrapper[4822]: I1124 14:51:54.987693 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62305997-e4f9-4e0e-9e28-0dea4382bf40-dns-svc\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.004258 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:51:55 crc kubenswrapper[4822]: E1124 14:51:55.004702 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerName="cloudkitty-api" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.004715 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerName="cloudkitty-api" Nov 24 14:51:55 crc kubenswrapper[4822]: E1124 14:51:55.004734 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerName="cloudkitty-api-log" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.004740 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerName="cloudkitty-api-log" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.004926 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerName="cloudkitty-api" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.004943 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" containerName="cloudkitty-api-log" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.005694 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.010514 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.025253 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.035008 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4kq9\" (UniqueName: \"kubernetes.io/projected/62305997-e4f9-4e0e-9e28-0dea4382bf40-kube-api-access-g4kq9\") pod \"dnsmasq-dns-79c5f76cb7-gj65d\" (UID: \"62305997-e4f9-4e0e-9e28-0dea4382bf40\") " pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.084096 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxx8s\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-kube-api-access-hxx8s\") pod \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.084259 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-certs\") pod \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.084325 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-combined-ca-bundle\") pod \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.084382 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-scripts\") pod \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.084409 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data\") pod \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.084445 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-internal-tls-certs\") pod \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.084472 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-logs\") pod \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.084506 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-public-tls-certs\") pod \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.084536 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data-custom\") pod \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\" (UID: \"51daf6fb-63aa-4ab7-8e3b-69e861e67e43\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.091123 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-logs" (OuterVolumeSpecName: "logs") pod "51daf6fb-63aa-4ab7-8e3b-69e861e67e43" (UID: "51daf6fb-63aa-4ab7-8e3b-69e861e67e43"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.092597 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-scripts" (OuterVolumeSpecName: "scripts") pod "51daf6fb-63aa-4ab7-8e3b-69e861e67e43" (UID: "51daf6fb-63aa-4ab7-8e3b-69e861e67e43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.094353 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "51daf6fb-63aa-4ab7-8e3b-69e861e67e43" (UID: "51daf6fb-63aa-4ab7-8e3b-69e861e67e43"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.096398 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-certs" (OuterVolumeSpecName: "certs") pod "51daf6fb-63aa-4ab7-8e3b-69e861e67e43" (UID: "51daf6fb-63aa-4ab7-8e3b-69e861e67e43"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.105610 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-kube-api-access-hxx8s" (OuterVolumeSpecName: "kube-api-access-hxx8s") pod "51daf6fb-63aa-4ab7-8e3b-69e861e67e43" (UID: "51daf6fb-63aa-4ab7-8e3b-69e861e67e43"). InnerVolumeSpecName "kube-api-access-hxx8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.131095 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51daf6fb-63aa-4ab7-8e3b-69e861e67e43" (UID: "51daf6fb-63aa-4ab7-8e3b-69e861e67e43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.132253 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data" (OuterVolumeSpecName: "config-data") pod "51daf6fb-63aa-4ab7-8e3b-69e861e67e43" (UID: "51daf6fb-63aa-4ab7-8e3b-69e861e67e43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.187909 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-scripts\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.187964 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188072 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e3e633d9-cd8a-4366-a09a-34c653b3275c-certs\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188088 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs98h\" (UniqueName: \"kubernetes.io/projected/e3e633d9-cd8a-4366-a09a-34c653b3275c-kube-api-access-zs98h\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188139 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188190 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-config-data\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188357 4822 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188370 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188380 4822 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188390 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188401 4822 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188411 4822 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.188425 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxx8s\" (UniqueName: \"kubernetes.io/projected/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-kube-api-access-hxx8s\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.194057 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "51daf6fb-63aa-4ab7-8e3b-69e861e67e43" (UID: "51daf6fb-63aa-4ab7-8e3b-69e861e67e43"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.203756 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "51daf6fb-63aa-4ab7-8e3b-69e861e67e43" (UID: "51daf6fb-63aa-4ab7-8e3b-69e861e67e43"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.240961 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.289718 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-scripts\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.289757 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.289905 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e3e633d9-cd8a-4366-a09a-34c653b3275c-certs\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.289924 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs98h\" (UniqueName: \"kubernetes.io/projected/e3e633d9-cd8a-4366-a09a-34c653b3275c-kube-api-access-zs98h\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.289965 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.290006 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-config-data\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.290068 4822 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.290079 4822 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51daf6fb-63aa-4ab7-8e3b-69e861e67e43-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.294911 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-config-data\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.295321 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.295733 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-scripts\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.296456 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e633d9-cd8a-4366-a09a-34c653b3275c-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.298091 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e3e633d9-cd8a-4366-a09a-34c653b3275c-certs\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.305445 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs98h\" (UniqueName: \"kubernetes.io/projected/e3e633d9-cd8a-4366-a09a-34c653b3275c-kube-api-access-zs98h\") pod \"cloudkitty-proc-0\" (UID: \"e3e633d9-cd8a-4366-a09a-34c653b3275c\") " pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.462650 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af242c06-68a2-436b-84d9-c2d020f15903","Type":"ContainerStarted","Data":"472b8ab81d0b7135a5694f1c5c085a65a81db5971a4de6108e790bb88bc46ed5"} Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.468042 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.468986 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.471728 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"51daf6fb-63aa-4ab7-8e3b-69e861e67e43","Type":"ContainerDied","Data":"d8dc33f4986b6160e6dccc5041abadad819e63c306e0c2fd51e4a684fcbb37e2"} Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.471780 4822 scope.go:117] "RemoveContainer" containerID="a8fe55b14cbb0e20e3997f9943574c2356792e3425cd3798ec3cc8bfa3a82050" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.479222 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.503708 4822 scope.go:117] "RemoveContainer" containerID="0ed5ee75ecb4ed8f0fbfe3dff427d71463d18ecc6ae018bce84cf9fcbe2de39b" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.529032 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.545052 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.549751 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.555290 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.556906 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.559570 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.559647 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.560622 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.593472 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.599632 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-sb\") pod \"b79568a5-e8f5-479c-9978-d55609cb5f59\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.599848 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bgmw\" (UniqueName: \"kubernetes.io/projected/b79568a5-e8f5-479c-9978-d55609cb5f59-kube-api-access-9bgmw\") pod \"b79568a5-e8f5-479c-9978-d55609cb5f59\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.599936 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-swift-storage-0\") pod \"b79568a5-e8f5-479c-9978-d55609cb5f59\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.607578 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-nb\") pod \"b79568a5-e8f5-479c-9978-d55609cb5f59\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.607698 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-openstack-edpm-ipam\") pod \"b79568a5-e8f5-479c-9978-d55609cb5f59\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.607825 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-svc\") pod \"b79568a5-e8f5-479c-9978-d55609cb5f59\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.607961 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-config\") pod \"b79568a5-e8f5-479c-9978-d55609cb5f59\" (UID: \"b79568a5-e8f5-479c-9978-d55609cb5f59\") " Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.600187 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b79568a5-e8f5-479c-9978-d55609cb5f59" (UID: "b79568a5-e8f5-479c-9978-d55609cb5f59"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.602558 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b79568a5-e8f5-479c-9978-d55609cb5f59" (UID: "b79568a5-e8f5-479c-9978-d55609cb5f59"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.610027 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b79568a5-e8f5-479c-9978-d55609cb5f59" (UID: "b79568a5-e8f5-479c-9978-d55609cb5f59"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.610361 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "b79568a5-e8f5-479c-9978-d55609cb5f59" (UID: "b79568a5-e8f5-479c-9978-d55609cb5f59"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.610536 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b79568a5-e8f5-479c-9978-d55609cb5f59" (UID: "b79568a5-e8f5-479c-9978-d55609cb5f59"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.610642 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-config" (OuterVolumeSpecName: "config") pod "b79568a5-e8f5-479c-9978-d55609cb5f59" (UID: "b79568a5-e8f5-479c-9978-d55609cb5f59"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.614499 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b79568a5-e8f5-479c-9978-d55609cb5f59-kube-api-access-9bgmw" (OuterVolumeSpecName: "kube-api-access-9bgmw") pod "b79568a5-e8f5-479c-9978-d55609cb5f59" (UID: "b79568a5-e8f5-479c-9978-d55609cb5f59"). InnerVolumeSpecName "kube-api-access-9bgmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714296 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714504 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714526 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714544 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-scripts\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714560 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr7wz\" (UniqueName: \"kubernetes.io/projected/4089c54f-b1cf-40c4-8369-872dabeb8202-kube-api-access-hr7wz\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714590 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-config-data\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714643 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714669 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4089c54f-b1cf-40c4-8369-872dabeb8202-logs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714704 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4089c54f-b1cf-40c4-8369-872dabeb8202-certs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714783 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714795 4822 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714819 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714827 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714835 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714846 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bgmw\" (UniqueName: \"kubernetes.io/projected/b79568a5-e8f5-479c-9978-d55609cb5f59-kube-api-access-9bgmw\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.714859 4822 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b79568a5-e8f5-479c-9978-d55609cb5f59-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.730771 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51daf6fb-63aa-4ab7-8e3b-69e861e67e43" path="/var/lib/kubelet/pods/51daf6fb-63aa-4ab7-8e3b-69e861e67e43/volumes" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.731494 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74794267-7a2e-4dc4-bb6b-35b99814e3ee" path="/var/lib/kubelet/pods/74794267-7a2e-4dc4-bb6b-35b99814e3ee/volumes" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.732139 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79c5f76cb7-gj65d"] Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.817665 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.817722 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.817751 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.817783 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-scripts\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.817798 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr7wz\" (UniqueName: \"kubernetes.io/projected/4089c54f-b1cf-40c4-8369-872dabeb8202-kube-api-access-hr7wz\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.817837 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-config-data\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.817882 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.817924 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4089c54f-b1cf-40c4-8369-872dabeb8202-logs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.817954 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4089c54f-b1cf-40c4-8369-872dabeb8202-certs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.826961 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4089c54f-b1cf-40c4-8369-872dabeb8202-logs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.829123 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.836705 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4089c54f-b1cf-40c4-8369-872dabeb8202-certs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.837045 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-scripts\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.838896 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-config-data\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.839854 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.841053 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.852486 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4089c54f-b1cf-40c4-8369-872dabeb8202-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.859775 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr7wz\" (UniqueName: \"kubernetes.io/projected/4089c54f-b1cf-40c4-8369-872dabeb8202-kube-api-access-hr7wz\") pod \"cloudkitty-api-0\" (UID: \"4089c54f-b1cf-40c4-8369-872dabeb8202\") " pod="openstack/cloudkitty-api-0" Nov 24 14:51:55 crc kubenswrapper[4822]: I1124 14:51:55.886200 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 14:51:56 crc kubenswrapper[4822]: I1124 14:51:56.239746 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 14:51:56 crc kubenswrapper[4822]: I1124 14:51:56.478516 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"e3e633d9-cd8a-4366-a09a-34c653b3275c","Type":"ContainerStarted","Data":"bfa9c9efcf8c5a12f49981745c271e0aa34bb34caca37e29572dff07043186c2"} Nov 24 14:51:56 crc kubenswrapper[4822]: I1124 14:51:56.487371 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" event={"ID":"62305997-e4f9-4e0e-9e28-0dea4382bf40","Type":"ContainerStarted","Data":"e303c1d2f53b6ca8a04e66dc205cd490543fbc1dd8ab2db0bb7bd6ebf4205129"} Nov 24 14:51:56 crc kubenswrapper[4822]: I1124 14:51:56.487525 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-584fd7dfd7-qdq77" Nov 24 14:51:56 crc kubenswrapper[4822]: I1124 14:51:56.557575 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 14:51:56 crc kubenswrapper[4822]: W1124 14:51:56.563515 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4089c54f_b1cf_40c4_8369_872dabeb8202.slice/crio-57280a91617d3fba5a5c2b756b0927d272aeacc0ff72abcace646e3db541d85e WatchSource:0}: Error finding container 57280a91617d3fba5a5c2b756b0927d272aeacc0ff72abcace646e3db541d85e: Status 404 returned error can't find the container with id 57280a91617d3fba5a5c2b756b0927d272aeacc0ff72abcace646e3db541d85e Nov 24 14:51:56 crc kubenswrapper[4822]: I1124 14:51:56.832611 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-584fd7dfd7-qdq77"] Nov 24 14:51:56 crc kubenswrapper[4822]: I1124 14:51:56.839634 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-584fd7dfd7-qdq77"] Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.511156 4822 generic.go:334] "Generic (PLEG): container finished" podID="62305997-e4f9-4e0e-9e28-0dea4382bf40" containerID="59cd2aee4d573ae008f71305f7fd8503ff0da6024722d74f7cd22aecb785eea9" exitCode=0 Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.511273 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" event={"ID":"62305997-e4f9-4e0e-9e28-0dea4382bf40","Type":"ContainerDied","Data":"59cd2aee4d573ae008f71305f7fd8503ff0da6024722d74f7cd22aecb785eea9"} Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.518563 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"4089c54f-b1cf-40c4-8369-872dabeb8202","Type":"ContainerStarted","Data":"17ccf4766cb142eb03c229974414d0aa8da8c30b94f2fd065c50935cacd5c68c"} Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.518617 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"4089c54f-b1cf-40c4-8369-872dabeb8202","Type":"ContainerStarted","Data":"165f863c04a7a2570fa2b6d6d042a2882367cdf5b45c59665525b9c7a9ea6502"} Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.518639 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"4089c54f-b1cf-40c4-8369-872dabeb8202","Type":"ContainerStarted","Data":"57280a91617d3fba5a5c2b756b0927d272aeacc0ff72abcace646e3db541d85e"} Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.520766 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"e3e633d9-cd8a-4366-a09a-34c653b3275c","Type":"ContainerStarted","Data":"22be0be5373f8f5e02b43ea1f848aa00bbdc722ec937e04c613669462c891a94"} Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.522158 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c969e517-0251-4537-9999-53bf0cfa7cf0","Type":"ContainerStarted","Data":"7002de1b21cc1cae3e9f377be65923b97de0aeaa78b7ba0df8dc42236244578b"} Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.525935 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.573486 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=3.328827786 podStartE2EDuration="3.573459039s" podCreationTimestamp="2025-11-24 14:51:54 +0000 UTC" firstStartedPulling="2025-11-24 14:51:56.253901752 +0000 UTC m=+1953.370542229" lastFinishedPulling="2025-11-24 14:51:56.498533005 +0000 UTC m=+1953.615173482" observedRunningTime="2025-11-24 14:51:57.565947841 +0000 UTC m=+1954.682588368" watchObservedRunningTime="2025-11-24 14:51:57.573459039 +0000 UTC m=+1954.690099546" Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.628098 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.628074669 podStartE2EDuration="2.628074669s" podCreationTimestamp="2025-11-24 14:51:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:51:57.620630903 +0000 UTC m=+1954.737271390" watchObservedRunningTime="2025-11-24 14:51:57.628074669 +0000 UTC m=+1954.744715146" Nov 24 14:51:57 crc kubenswrapper[4822]: I1124 14:51:57.718022 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b79568a5-e8f5-479c-9978-d55609cb5f59" path="/var/lib/kubelet/pods/b79568a5-e8f5-479c-9978-d55609cb5f59/volumes" Nov 24 14:51:58 crc kubenswrapper[4822]: I1124 14:51:58.534621 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" event={"ID":"62305997-e4f9-4e0e-9e28-0dea4382bf40","Type":"ContainerStarted","Data":"807eb3876d8894e621d31f3e5873699ef32e47bee1b2e48e6a60b840832a1d2a"} Nov 24 14:51:58 crc kubenswrapper[4822]: I1124 14:51:58.551926 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" podStartSLOduration=4.5518911840000005 podStartE2EDuration="4.551891184s" podCreationTimestamp="2025-11-24 14:51:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:51:58.551505883 +0000 UTC m=+1955.668146370" watchObservedRunningTime="2025-11-24 14:51:58.551891184 +0000 UTC m=+1955.668531661" Nov 24 14:51:59 crc kubenswrapper[4822]: I1124 14:51:59.545881 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:52:05 crc kubenswrapper[4822]: I1124 14:52:05.242450 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79c5f76cb7-gj65d" Nov 24 14:52:05 crc kubenswrapper[4822]: I1124 14:52:05.351266 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bf74d94c7-kp4rh"] Nov 24 14:52:05 crc kubenswrapper[4822]: I1124 14:52:05.351819 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" podUID="3d3e4572-1875-459c-9fbd-53ea02e8e123" containerName="dnsmasq-dns" containerID="cri-o://e8868dda0ec2298907ee7e1a61984e27b8adf12cdc09cf103e325220b3df05ab" gracePeriod=10 Nov 24 14:52:05 crc kubenswrapper[4822]: I1124 14:52:05.607357 4822 generic.go:334] "Generic (PLEG): container finished" podID="3d3e4572-1875-459c-9fbd-53ea02e8e123" containerID="e8868dda0ec2298907ee7e1a61984e27b8adf12cdc09cf103e325220b3df05ab" exitCode=0 Nov 24 14:52:05 crc kubenswrapper[4822]: I1124 14:52:05.607404 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" event={"ID":"3d3e4572-1875-459c-9fbd-53ea02e8e123","Type":"ContainerDied","Data":"e8868dda0ec2298907ee7e1a61984e27b8adf12cdc09cf103e325220b3df05ab"} Nov 24 14:52:05 crc kubenswrapper[4822]: I1124 14:52:05.989807 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.166725 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-svc\") pod \"3d3e4572-1875-459c-9fbd-53ea02e8e123\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.166772 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-nb\") pod \"3d3e4572-1875-459c-9fbd-53ea02e8e123\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.166808 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-swift-storage-0\") pod \"3d3e4572-1875-459c-9fbd-53ea02e8e123\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.166850 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-sb\") pod \"3d3e4572-1875-459c-9fbd-53ea02e8e123\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.166973 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzqft\" (UniqueName: \"kubernetes.io/projected/3d3e4572-1875-459c-9fbd-53ea02e8e123-kube-api-access-gzqft\") pod \"3d3e4572-1875-459c-9fbd-53ea02e8e123\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.166997 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-config\") pod \"3d3e4572-1875-459c-9fbd-53ea02e8e123\" (UID: \"3d3e4572-1875-459c-9fbd-53ea02e8e123\") " Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.187329 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d3e4572-1875-459c-9fbd-53ea02e8e123-kube-api-access-gzqft" (OuterVolumeSpecName: "kube-api-access-gzqft") pod "3d3e4572-1875-459c-9fbd-53ea02e8e123" (UID: "3d3e4572-1875-459c-9fbd-53ea02e8e123"). InnerVolumeSpecName "kube-api-access-gzqft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.237264 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-config" (OuterVolumeSpecName: "config") pod "3d3e4572-1875-459c-9fbd-53ea02e8e123" (UID: "3d3e4572-1875-459c-9fbd-53ea02e8e123"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.243372 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3d3e4572-1875-459c-9fbd-53ea02e8e123" (UID: "3d3e4572-1875-459c-9fbd-53ea02e8e123"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.244566 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3d3e4572-1875-459c-9fbd-53ea02e8e123" (UID: "3d3e4572-1875-459c-9fbd-53ea02e8e123"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.254736 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3d3e4572-1875-459c-9fbd-53ea02e8e123" (UID: "3d3e4572-1875-459c-9fbd-53ea02e8e123"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.255888 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3d3e4572-1875-459c-9fbd-53ea02e8e123" (UID: "3d3e4572-1875-459c-9fbd-53ea02e8e123"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.271293 4822 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.271340 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.271354 4822 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.271366 4822 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.271438 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzqft\" (UniqueName: \"kubernetes.io/projected/3d3e4572-1875-459c-9fbd-53ea02e8e123-kube-api-access-gzqft\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.271451 4822 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d3e4572-1875-459c-9fbd-53ea02e8e123-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.616250 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" event={"ID":"3d3e4572-1875-459c-9fbd-53ea02e8e123","Type":"ContainerDied","Data":"c6e709c36f31e650c64f829dbb3f898f0ea02428e630f822943f34b017906f42"} Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.616297 4822 scope.go:117] "RemoveContainer" containerID="e8868dda0ec2298907ee7e1a61984e27b8adf12cdc09cf103e325220b3df05ab" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.616404 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf74d94c7-kp4rh" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.674015 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bf74d94c7-kp4rh"] Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.682381 4822 scope.go:117] "RemoveContainer" containerID="6d8d8b424df4d1487c6a9f1bc2c3698433cee66cfe8301fd0e9184786fba1640" Nov 24 14:52:06 crc kubenswrapper[4822]: I1124 14:52:06.686747 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bf74d94c7-kp4rh"] Nov 24 14:52:06 crc kubenswrapper[4822]: E1124 14:52:06.811447 4822 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d3e4572_1875_459c_9fbd_53ea02e8e123.slice/crio-c6e709c36f31e650c64f829dbb3f898f0ea02428e630f822943f34b017906f42\": RecentStats: unable to find data in memory cache]" Nov 24 14:52:07 crc kubenswrapper[4822]: I1124 14:52:07.752110 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d3e4572-1875-459c-9fbd-53ea02e8e123" path="/var/lib/kubelet/pods/3d3e4572-1875-459c-9fbd-53ea02e8e123/volumes" Nov 24 14:52:13 crc kubenswrapper[4822]: I1124 14:52:13.868535 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.238057 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d"] Nov 24 14:52:14 crc kubenswrapper[4822]: E1124 14:52:14.238516 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d3e4572-1875-459c-9fbd-53ea02e8e123" containerName="init" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.238541 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d3e4572-1875-459c-9fbd-53ea02e8e123" containerName="init" Nov 24 14:52:14 crc kubenswrapper[4822]: E1124 14:52:14.238567 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d3e4572-1875-459c-9fbd-53ea02e8e123" containerName="dnsmasq-dns" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.238575 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d3e4572-1875-459c-9fbd-53ea02e8e123" containerName="dnsmasq-dns" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.238827 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d3e4572-1875-459c-9fbd-53ea02e8e123" containerName="dnsmasq-dns" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.239788 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.243485 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.243613 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.243886 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.243896 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.253745 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d"] Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.341611 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.341687 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.342018 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.342087 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cggg\" (UniqueName: \"kubernetes.io/projected/4b17a1c2-3060-401f-89a2-21f6e7edec54-kube-api-access-2cggg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.444525 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.444840 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cggg\" (UniqueName: \"kubernetes.io/projected/4b17a1c2-3060-401f-89a2-21f6e7edec54-kube-api-access-2cggg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.445010 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.445075 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.450295 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.450803 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.453127 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.470056 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cggg\" (UniqueName: \"kubernetes.io/projected/4b17a1c2-3060-401f-89a2-21f6e7edec54-kube-api-access-2cggg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:14 crc kubenswrapper[4822]: I1124 14:52:14.571265 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:15 crc kubenswrapper[4822]: W1124 14:52:15.259675 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b17a1c2_3060_401f_89a2_21f6e7edec54.slice/crio-0b0e2351652834d5566ae1d47de6b1b959465bde647c243a98a873e3721ff4b9 WatchSource:0}: Error finding container 0b0e2351652834d5566ae1d47de6b1b959465bde647c243a98a873e3721ff4b9: Status 404 returned error can't find the container with id 0b0e2351652834d5566ae1d47de6b1b959465bde647c243a98a873e3721ff4b9 Nov 24 14:52:15 crc kubenswrapper[4822]: I1124 14:52:15.264888 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d"] Nov 24 14:52:15 crc kubenswrapper[4822]: I1124 14:52:15.720572 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" event={"ID":"4b17a1c2-3060-401f-89a2-21f6e7edec54","Type":"ContainerStarted","Data":"0b0e2351652834d5566ae1d47de6b1b959465bde647c243a98a873e3721ff4b9"} Nov 24 14:52:24 crc kubenswrapper[4822]: I1124 14:52:24.520356 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:52:24 crc kubenswrapper[4822]: I1124 14:52:24.812928 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" event={"ID":"4b17a1c2-3060-401f-89a2-21f6e7edec54","Type":"ContainerStarted","Data":"3175942c8126cf6e2f72ee43f970e5f1638ebc0f802240bde37f076d44e6f987"} Nov 24 14:52:24 crc kubenswrapper[4822]: I1124 14:52:24.843322 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" podStartSLOduration=1.5909884239999998 podStartE2EDuration="10.843296705s" podCreationTimestamp="2025-11-24 14:52:14 +0000 UTC" firstStartedPulling="2025-11-24 14:52:15.263688572 +0000 UTC m=+1972.380329059" lastFinishedPulling="2025-11-24 14:52:24.515996863 +0000 UTC m=+1981.632637340" observedRunningTime="2025-11-24 14:52:24.831002446 +0000 UTC m=+1981.947642953" watchObservedRunningTime="2025-11-24 14:52:24.843296705 +0000 UTC m=+1981.959937222" Nov 24 14:52:27 crc kubenswrapper[4822]: I1124 14:52:27.845320 4822 generic.go:334] "Generic (PLEG): container finished" podID="af242c06-68a2-436b-84d9-c2d020f15903" containerID="472b8ab81d0b7135a5694f1c5c085a65a81db5971a4de6108e790bb88bc46ed5" exitCode=0 Nov 24 14:52:27 crc kubenswrapper[4822]: I1124 14:52:27.845344 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af242c06-68a2-436b-84d9-c2d020f15903","Type":"ContainerDied","Data":"472b8ab81d0b7135a5694f1c5c085a65a81db5971a4de6108e790bb88bc46ed5"} Nov 24 14:52:28 crc kubenswrapper[4822]: I1124 14:52:28.859125 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af242c06-68a2-436b-84d9-c2d020f15903","Type":"ContainerStarted","Data":"6c90b2393f5a1582af5ca7d2b0288850f43258ca8dfd1c6b740de3f0be95e911"} Nov 24 14:52:28 crc kubenswrapper[4822]: I1124 14:52:28.859930 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 14:52:28 crc kubenswrapper[4822]: I1124 14:52:28.909190 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.90916199 podStartE2EDuration="36.90916199s" podCreationTimestamp="2025-11-24 14:51:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:52:28.894710792 +0000 UTC m=+1986.011351289" watchObservedRunningTime="2025-11-24 14:52:28.90916199 +0000 UTC m=+1986.025802477" Nov 24 14:52:29 crc kubenswrapper[4822]: I1124 14:52:29.874181 4822 generic.go:334] "Generic (PLEG): container finished" podID="c969e517-0251-4537-9999-53bf0cfa7cf0" containerID="7002de1b21cc1cae3e9f377be65923b97de0aeaa78b7ba0df8dc42236244578b" exitCode=0 Nov 24 14:52:29 crc kubenswrapper[4822]: I1124 14:52:29.874307 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c969e517-0251-4537-9999-53bf0cfa7cf0","Type":"ContainerDied","Data":"7002de1b21cc1cae3e9f377be65923b97de0aeaa78b7ba0df8dc42236244578b"} Nov 24 14:52:30 crc kubenswrapper[4822]: I1124 14:52:30.888537 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c969e517-0251-4537-9999-53bf0cfa7cf0","Type":"ContainerStarted","Data":"63507aa76f39b22bc9da9c33082a0df117a0fffed9f87d58009ce4d8a148a649"} Nov 24 14:52:30 crc kubenswrapper[4822]: I1124 14:52:30.890606 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:52:30 crc kubenswrapper[4822]: I1124 14:52:30.933762 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.933735288 podStartE2EDuration="37.933735288s" podCreationTimestamp="2025-11-24 14:51:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:52:30.92843427 +0000 UTC m=+1988.045074757" watchObservedRunningTime="2025-11-24 14:52:30.933735288 +0000 UTC m=+1988.050375765" Nov 24 14:52:32 crc kubenswrapper[4822]: I1124 14:52:32.683930 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Nov 24 14:52:36 crc kubenswrapper[4822]: I1124 14:52:36.958152 4822 generic.go:334] "Generic (PLEG): container finished" podID="4b17a1c2-3060-401f-89a2-21f6e7edec54" containerID="3175942c8126cf6e2f72ee43f970e5f1638ebc0f802240bde37f076d44e6f987" exitCode=0 Nov 24 14:52:36 crc kubenswrapper[4822]: I1124 14:52:36.958248 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" event={"ID":"4b17a1c2-3060-401f-89a2-21f6e7edec54","Type":"ContainerDied","Data":"3175942c8126cf6e2f72ee43f970e5f1638ebc0f802240bde37f076d44e6f987"} Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.515730 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.609299 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-ssh-key\") pod \"4b17a1c2-3060-401f-89a2-21f6e7edec54\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.609409 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-inventory\") pod \"4b17a1c2-3060-401f-89a2-21f6e7edec54\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.609574 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cggg\" (UniqueName: \"kubernetes.io/projected/4b17a1c2-3060-401f-89a2-21f6e7edec54-kube-api-access-2cggg\") pod \"4b17a1c2-3060-401f-89a2-21f6e7edec54\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.609658 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-repo-setup-combined-ca-bundle\") pod \"4b17a1c2-3060-401f-89a2-21f6e7edec54\" (UID: \"4b17a1c2-3060-401f-89a2-21f6e7edec54\") " Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.616439 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b17a1c2-3060-401f-89a2-21f6e7edec54-kube-api-access-2cggg" (OuterVolumeSpecName: "kube-api-access-2cggg") pod "4b17a1c2-3060-401f-89a2-21f6e7edec54" (UID: "4b17a1c2-3060-401f-89a2-21f6e7edec54"). InnerVolumeSpecName "kube-api-access-2cggg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.618206 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4b17a1c2-3060-401f-89a2-21f6e7edec54" (UID: "4b17a1c2-3060-401f-89a2-21f6e7edec54"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.639488 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-inventory" (OuterVolumeSpecName: "inventory") pod "4b17a1c2-3060-401f-89a2-21f6e7edec54" (UID: "4b17a1c2-3060-401f-89a2-21f6e7edec54"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.663014 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b17a1c2-3060-401f-89a2-21f6e7edec54" (UID: "4b17a1c2-3060-401f-89a2-21f6e7edec54"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.713215 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cggg\" (UniqueName: \"kubernetes.io/projected/4b17a1c2-3060-401f-89a2-21f6e7edec54-kube-api-access-2cggg\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.713365 4822 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.713386 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.713406 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b17a1c2-3060-401f-89a2-21f6e7edec54-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.989604 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" event={"ID":"4b17a1c2-3060-401f-89a2-21f6e7edec54","Type":"ContainerDied","Data":"0b0e2351652834d5566ae1d47de6b1b959465bde647c243a98a873e3721ff4b9"} Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.990144 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b0e2351652834d5566ae1d47de6b1b959465bde647c243a98a873e3721ff4b9" Nov 24 14:52:38 crc kubenswrapper[4822]: I1124 14:52:38.989724 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.141887 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w"] Nov 24 14:52:39 crc kubenswrapper[4822]: E1124 14:52:39.142614 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b17a1c2-3060-401f-89a2-21f6e7edec54" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.142647 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b17a1c2-3060-401f-89a2-21f6e7edec54" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.143015 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b17a1c2-3060-401f-89a2-21f6e7edec54" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.144315 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.146872 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.147414 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.147883 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.149999 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.159617 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w"] Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.231067 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh8pt\" (UniqueName: \"kubernetes.io/projected/0fc7b81a-075b-4f00-b4b1-d9863317c650-kube-api-access-fh8pt\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dx95w\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.231705 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dx95w\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.231931 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dx95w\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.333936 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dx95w\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.334051 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dx95w\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.334168 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh8pt\" (UniqueName: \"kubernetes.io/projected/0fc7b81a-075b-4f00-b4b1-d9863317c650-kube-api-access-fh8pt\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dx95w\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.338476 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dx95w\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.339000 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dx95w\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.350509 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh8pt\" (UniqueName: \"kubernetes.io/projected/0fc7b81a-075b-4f00-b4b1-d9863317c650-kube-api-access-fh8pt\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dx95w\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:39 crc kubenswrapper[4822]: I1124 14:52:39.531085 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:40 crc kubenswrapper[4822]: I1124 14:52:40.104650 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w"] Nov 24 14:52:41 crc kubenswrapper[4822]: I1124 14:52:41.015499 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" event={"ID":"0fc7b81a-075b-4f00-b4b1-d9863317c650","Type":"ContainerStarted","Data":"44cfa48dc617f455e9236f642ab96812a403b0f129b8c47acd68acadb7697d9a"} Nov 24 14:52:41 crc kubenswrapper[4822]: I1124 14:52:41.016060 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" event={"ID":"0fc7b81a-075b-4f00-b4b1-d9863317c650","Type":"ContainerStarted","Data":"0e0134ebe5363a2d5f2d2529e9aaf862b397b780f9c7cac1e7726b8463e0fc21"} Nov 24 14:52:41 crc kubenswrapper[4822]: I1124 14:52:41.043899 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" podStartSLOduration=1.591491246 podStartE2EDuration="2.043859001s" podCreationTimestamp="2025-11-24 14:52:39 +0000 UTC" firstStartedPulling="2025-11-24 14:52:40.114489751 +0000 UTC m=+1997.231130268" lastFinishedPulling="2025-11-24 14:52:40.566857536 +0000 UTC m=+1997.683498023" observedRunningTime="2025-11-24 14:52:41.032255794 +0000 UTC m=+1998.148896331" watchObservedRunningTime="2025-11-24 14:52:41.043859001 +0000 UTC m=+1998.160499488" Nov 24 14:52:41 crc kubenswrapper[4822]: I1124 14:52:41.578487 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:52:41 crc kubenswrapper[4822]: I1124 14:52:41.578599 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:52:42 crc kubenswrapper[4822]: I1124 14:52:42.925721 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 14:52:43 crc kubenswrapper[4822]: I1124 14:52:43.860394 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:52:44 crc kubenswrapper[4822]: I1124 14:52:44.050071 4822 generic.go:334] "Generic (PLEG): container finished" podID="0fc7b81a-075b-4f00-b4b1-d9863317c650" containerID="44cfa48dc617f455e9236f642ab96812a403b0f129b8c47acd68acadb7697d9a" exitCode=0 Nov 24 14:52:44 crc kubenswrapper[4822]: I1124 14:52:44.050113 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" event={"ID":"0fc7b81a-075b-4f00-b4b1-d9863317c650","Type":"ContainerDied","Data":"44cfa48dc617f455e9236f642ab96812a403b0f129b8c47acd68acadb7697d9a"} Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.622111 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.694494 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-inventory\") pod \"0fc7b81a-075b-4f00-b4b1-d9863317c650\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.694556 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-ssh-key\") pod \"0fc7b81a-075b-4f00-b4b1-d9863317c650\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.694671 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh8pt\" (UniqueName: \"kubernetes.io/projected/0fc7b81a-075b-4f00-b4b1-d9863317c650-kube-api-access-fh8pt\") pod \"0fc7b81a-075b-4f00-b4b1-d9863317c650\" (UID: \"0fc7b81a-075b-4f00-b4b1-d9863317c650\") " Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.716224 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fc7b81a-075b-4f00-b4b1-d9863317c650-kube-api-access-fh8pt" (OuterVolumeSpecName: "kube-api-access-fh8pt") pod "0fc7b81a-075b-4f00-b4b1-d9863317c650" (UID: "0fc7b81a-075b-4f00-b4b1-d9863317c650"). InnerVolumeSpecName "kube-api-access-fh8pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.726421 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-inventory" (OuterVolumeSpecName: "inventory") pod "0fc7b81a-075b-4f00-b4b1-d9863317c650" (UID: "0fc7b81a-075b-4f00-b4b1-d9863317c650"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.736305 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0fc7b81a-075b-4f00-b4b1-d9863317c650" (UID: "0fc7b81a-075b-4f00-b4b1-d9863317c650"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.798801 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.798967 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh8pt\" (UniqueName: \"kubernetes.io/projected/0fc7b81a-075b-4f00-b4b1-d9863317c650-kube-api-access-fh8pt\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:45 crc kubenswrapper[4822]: I1124 14:52:45.799050 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fc7b81a-075b-4f00-b4b1-d9863317c650-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.070822 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" event={"ID":"0fc7b81a-075b-4f00-b4b1-d9863317c650","Type":"ContainerDied","Data":"0e0134ebe5363a2d5f2d2529e9aaf862b397b780f9c7cac1e7726b8463e0fc21"} Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.070856 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e0134ebe5363a2d5f2d2529e9aaf862b397b780f9c7cac1e7726b8463e0fc21" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.070884 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dx95w" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.144435 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk"] Nov 24 14:52:46 crc kubenswrapper[4822]: E1124 14:52:46.145136 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fc7b81a-075b-4f00-b4b1-d9863317c650" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.145154 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fc7b81a-075b-4f00-b4b1-d9863317c650" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.145417 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fc7b81a-075b-4f00-b4b1-d9863317c650" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.146129 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.149397 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.149642 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.149585 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.152193 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.156730 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk"] Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.207651 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lwzn\" (UniqueName: \"kubernetes.io/projected/3cea7046-1628-413a-99d0-aa36783ec7fc-kube-api-access-6lwzn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.207950 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.208078 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.208233 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.310469 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.310749 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.311002 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lwzn\" (UniqueName: \"kubernetes.io/projected/3cea7046-1628-413a-99d0-aa36783ec7fc-kube-api-access-6lwzn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.311152 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.322124 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.325808 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.336946 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.340161 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lwzn\" (UniqueName: \"kubernetes.io/projected/3cea7046-1628-413a-99d0-aa36783ec7fc-kube-api-access-6lwzn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:46 crc kubenswrapper[4822]: I1124 14:52:46.485003 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:52:47 crc kubenswrapper[4822]: I1124 14:52:47.085828 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk"] Nov 24 14:52:47 crc kubenswrapper[4822]: W1124 14:52:47.092444 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cea7046_1628_413a_99d0_aa36783ec7fc.slice/crio-b9083e946b1fcb6f85c91a0711bf2e9f6ac9a9abe16e8067f0e9a9d6ab53ae06 WatchSource:0}: Error finding container b9083e946b1fcb6f85c91a0711bf2e9f6ac9a9abe16e8067f0e9a9d6ab53ae06: Status 404 returned error can't find the container with id b9083e946b1fcb6f85c91a0711bf2e9f6ac9a9abe16e8067f0e9a9d6ab53ae06 Nov 24 14:52:48 crc kubenswrapper[4822]: I1124 14:52:48.097338 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" event={"ID":"3cea7046-1628-413a-99d0-aa36783ec7fc","Type":"ContainerStarted","Data":"05e1ca3e885f17ba77ab4700f973ff9b8fbe9382da72903ede36ca92f7ce2964"} Nov 24 14:52:48 crc kubenswrapper[4822]: I1124 14:52:48.097696 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" event={"ID":"3cea7046-1628-413a-99d0-aa36783ec7fc","Type":"ContainerStarted","Data":"b9083e946b1fcb6f85c91a0711bf2e9f6ac9a9abe16e8067f0e9a9d6ab53ae06"} Nov 24 14:52:48 crc kubenswrapper[4822]: I1124 14:52:48.122254 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" podStartSLOduration=1.689321884 podStartE2EDuration="2.122234893s" podCreationTimestamp="2025-11-24 14:52:46 +0000 UTC" firstStartedPulling="2025-11-24 14:52:47.095421432 +0000 UTC m=+2004.212061939" lastFinishedPulling="2025-11-24 14:52:47.528334411 +0000 UTC m=+2004.644974948" observedRunningTime="2025-11-24 14:52:48.109478528 +0000 UTC m=+2005.226119025" watchObservedRunningTime="2025-11-24 14:52:48.122234893 +0000 UTC m=+2005.238875370" Nov 24 14:53:11 crc kubenswrapper[4822]: I1124 14:53:11.578264 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:53:11 crc kubenswrapper[4822]: I1124 14:53:11.578932 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:53:41 crc kubenswrapper[4822]: I1124 14:53:41.578687 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:53:41 crc kubenswrapper[4822]: I1124 14:53:41.579369 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:53:41 crc kubenswrapper[4822]: I1124 14:53:41.579424 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:53:41 crc kubenswrapper[4822]: I1124 14:53:41.580359 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5cc0b1343a16a8d188981e3c048d5b5d39678fa534e29571730f78b864e29cf8"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:53:41 crc kubenswrapper[4822]: I1124 14:53:41.580456 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://5cc0b1343a16a8d188981e3c048d5b5d39678fa534e29571730f78b864e29cf8" gracePeriod=600 Nov 24 14:53:41 crc kubenswrapper[4822]: I1124 14:53:41.781142 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="5cc0b1343a16a8d188981e3c048d5b5d39678fa534e29571730f78b864e29cf8" exitCode=0 Nov 24 14:53:41 crc kubenswrapper[4822]: I1124 14:53:41.781193 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"5cc0b1343a16a8d188981e3c048d5b5d39678fa534e29571730f78b864e29cf8"} Nov 24 14:53:41 crc kubenswrapper[4822]: I1124 14:53:41.781249 4822 scope.go:117] "RemoveContainer" containerID="7b2835acec527d77efba303bd1418ca3ce23bae040cebafe6ba650466a593678" Nov 24 14:53:42 crc kubenswrapper[4822]: I1124 14:53:42.795966 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300"} Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.021852 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tq6m5"] Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.025866 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.051792 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tq6m5"] Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.120327 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e-utilities\") pod \"community-operators-tq6m5\" (UID: \"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e\") " pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.120460 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d659v\" (UniqueName: \"kubernetes.io/projected/fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e-kube-api-access-d659v\") pod \"community-operators-tq6m5\" (UID: \"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e\") " pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.120565 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e-catalog-content\") pod \"community-operators-tq6m5\" (UID: \"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e\") " pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.221887 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e-utilities\") pod \"community-operators-tq6m5\" (UID: \"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e\") " pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.221986 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d659v\" (UniqueName: \"kubernetes.io/projected/fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e-kube-api-access-d659v\") pod \"community-operators-tq6m5\" (UID: \"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e\") " pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.222086 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e-catalog-content\") pod \"community-operators-tq6m5\" (UID: \"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e\") " pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.222493 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e-utilities\") pod \"community-operators-tq6m5\" (UID: \"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e\") " pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.222528 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e-catalog-content\") pod \"community-operators-tq6m5\" (UID: \"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e\") " pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.250635 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d659v\" (UniqueName: \"kubernetes.io/projected/fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e-kube-api-access-d659v\") pod \"community-operators-tq6m5\" (UID: \"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e\") " pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:44 crc kubenswrapper[4822]: I1124 14:53:44.385665 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:45 crc kubenswrapper[4822]: I1124 14:53:45.133860 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tq6m5"] Nov 24 14:53:45 crc kubenswrapper[4822]: I1124 14:53:45.826440 4822 generic.go:334] "Generic (PLEG): container finished" podID="fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e" containerID="a02d5228cad0583d4a5d3c67e5b538c23bec0748cd24c832c3d4aaed6450a8de" exitCode=0 Nov 24 14:53:45 crc kubenswrapper[4822]: I1124 14:53:45.826502 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6m5" event={"ID":"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e","Type":"ContainerDied","Data":"a02d5228cad0583d4a5d3c67e5b538c23bec0748cd24c832c3d4aaed6450a8de"} Nov 24 14:53:45 crc kubenswrapper[4822]: I1124 14:53:45.826755 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6m5" event={"ID":"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e","Type":"ContainerStarted","Data":"7df3dcea60108775aa0f629bd229e0768cb6e8b1455f2cd7094c1eae4f611530"} Nov 24 14:53:50 crc kubenswrapper[4822]: I1124 14:53:50.881003 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6m5" event={"ID":"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e","Type":"ContainerStarted","Data":"94a9b3416f68239498a624dc757ed682c32d82760aee4a9da77445c3b3544d4c"} Nov 24 14:53:51 crc kubenswrapper[4822]: I1124 14:53:51.894888 4822 generic.go:334] "Generic (PLEG): container finished" podID="fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e" containerID="94a9b3416f68239498a624dc757ed682c32d82760aee4a9da77445c3b3544d4c" exitCode=0 Nov 24 14:53:51 crc kubenswrapper[4822]: I1124 14:53:51.894946 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6m5" event={"ID":"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e","Type":"ContainerDied","Data":"94a9b3416f68239498a624dc757ed682c32d82760aee4a9da77445c3b3544d4c"} Nov 24 14:53:52 crc kubenswrapper[4822]: I1124 14:53:52.907894 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6m5" event={"ID":"fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e","Type":"ContainerStarted","Data":"39fe91f519253bee0f9817c399bc0b856f6708b027fee671a0ade951fa0465af"} Nov 24 14:53:52 crc kubenswrapper[4822]: I1124 14:53:52.937136 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tq6m5" podStartSLOduration=3.479052564 podStartE2EDuration="9.937121366s" podCreationTimestamp="2025-11-24 14:53:43 +0000 UTC" firstStartedPulling="2025-11-24 14:53:45.828672753 +0000 UTC m=+2062.945313270" lastFinishedPulling="2025-11-24 14:53:52.286741555 +0000 UTC m=+2069.403382072" observedRunningTime="2025-11-24 14:53:52.934469101 +0000 UTC m=+2070.051109608" watchObservedRunningTime="2025-11-24 14:53:52.937121366 +0000 UTC m=+2070.053761843" Nov 24 14:53:54 crc kubenswrapper[4822]: I1124 14:53:54.385890 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:54 crc kubenswrapper[4822]: I1124 14:53:54.386311 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:53:55 crc kubenswrapper[4822]: I1124 14:53:55.454564 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tq6m5" podUID="fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e" containerName="registry-server" probeResult="failure" output=< Nov 24 14:53:55 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:53:55 crc kubenswrapper[4822]: > Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.527903 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rzcn5"] Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.532796 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.545965 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rzcn5"] Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.553338 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-catalog-content\") pod \"certified-operators-rzcn5\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.553447 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgqhr\" (UniqueName: \"kubernetes.io/projected/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-kube-api-access-tgqhr\") pod \"certified-operators-rzcn5\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.553628 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-utilities\") pod \"certified-operators-rzcn5\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.655448 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgqhr\" (UniqueName: \"kubernetes.io/projected/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-kube-api-access-tgqhr\") pod \"certified-operators-rzcn5\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.655570 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-utilities\") pod \"certified-operators-rzcn5\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.655651 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-catalog-content\") pod \"certified-operators-rzcn5\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.656182 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-catalog-content\") pod \"certified-operators-rzcn5\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.656629 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-utilities\") pod \"certified-operators-rzcn5\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.686508 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgqhr\" (UniqueName: \"kubernetes.io/projected/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-kube-api-access-tgqhr\") pod \"certified-operators-rzcn5\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:02 crc kubenswrapper[4822]: I1124 14:54:02.864982 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:03 crc kubenswrapper[4822]: I1124 14:54:03.504659 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rzcn5"] Nov 24 14:54:04 crc kubenswrapper[4822]: I1124 14:54:04.054654 4822 generic.go:334] "Generic (PLEG): container finished" podID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerID="300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb" exitCode=0 Nov 24 14:54:04 crc kubenswrapper[4822]: I1124 14:54:04.054743 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzcn5" event={"ID":"c8c81345-38a8-4545-abeb-b6e1c2fc2c87","Type":"ContainerDied","Data":"300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb"} Nov 24 14:54:04 crc kubenswrapper[4822]: I1124 14:54:04.055130 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzcn5" event={"ID":"c8c81345-38a8-4545-abeb-b6e1c2fc2c87","Type":"ContainerStarted","Data":"ba91836fae5d1662da9b797c3e72f93707a63c1ac11574e4e57927068e292b2d"} Nov 24 14:54:04 crc kubenswrapper[4822]: I1124 14:54:04.465704 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:54:04 crc kubenswrapper[4822]: I1124 14:54:04.565005 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tq6m5" Nov 24 14:54:05 crc kubenswrapper[4822]: I1124 14:54:05.065613 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzcn5" event={"ID":"c8c81345-38a8-4545-abeb-b6e1c2fc2c87","Type":"ContainerStarted","Data":"e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f"} Nov 24 14:54:06 crc kubenswrapper[4822]: I1124 14:54:06.551004 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tq6m5"] Nov 24 14:54:06 crc kubenswrapper[4822]: I1124 14:54:06.902579 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9mzfn"] Nov 24 14:54:06 crc kubenswrapper[4822]: I1124 14:54:06.902812 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9mzfn" podUID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerName="registry-server" containerID="cri-o://822a597b520e7a18214ee4487b78524bac10e3725fb3a0d2795a6cf5f81a9eaf" gracePeriod=2 Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.085930 4822 generic.go:334] "Generic (PLEG): container finished" podID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerID="e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f" exitCode=0 Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.085994 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzcn5" event={"ID":"c8c81345-38a8-4545-abeb-b6e1c2fc2c87","Type":"ContainerDied","Data":"e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f"} Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.088441 4822 generic.go:334] "Generic (PLEG): container finished" podID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerID="822a597b520e7a18214ee4487b78524bac10e3725fb3a0d2795a6cf5f81a9eaf" exitCode=0 Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.088469 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mzfn" event={"ID":"d5e9ae7c-8cf5-404a-8403-94c5005ade7e","Type":"ContainerDied","Data":"822a597b520e7a18214ee4487b78524bac10e3725fb3a0d2795a6cf5f81a9eaf"} Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.437000 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.567488 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwkw5\" (UniqueName: \"kubernetes.io/projected/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-kube-api-access-mwkw5\") pod \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.567731 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-catalog-content\") pod \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.567774 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-utilities\") pod \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\" (UID: \"d5e9ae7c-8cf5-404a-8403-94c5005ade7e\") " Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.570504 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-utilities" (OuterVolumeSpecName: "utilities") pod "d5e9ae7c-8cf5-404a-8403-94c5005ade7e" (UID: "d5e9ae7c-8cf5-404a-8403-94c5005ade7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.574373 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-kube-api-access-mwkw5" (OuterVolumeSpecName: "kube-api-access-mwkw5") pod "d5e9ae7c-8cf5-404a-8403-94c5005ade7e" (UID: "d5e9ae7c-8cf5-404a-8403-94c5005ade7e"). InnerVolumeSpecName "kube-api-access-mwkw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.648045 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5e9ae7c-8cf5-404a-8403-94c5005ade7e" (UID: "d5e9ae7c-8cf5-404a-8403-94c5005ade7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.669981 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.670007 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:07 crc kubenswrapper[4822]: I1124 14:54:07.670017 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwkw5\" (UniqueName: \"kubernetes.io/projected/d5e9ae7c-8cf5-404a-8403-94c5005ade7e-kube-api-access-mwkw5\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:08 crc kubenswrapper[4822]: I1124 14:54:08.103620 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzcn5" event={"ID":"c8c81345-38a8-4545-abeb-b6e1c2fc2c87","Type":"ContainerStarted","Data":"c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0"} Nov 24 14:54:08 crc kubenswrapper[4822]: I1124 14:54:08.106830 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mzfn" event={"ID":"d5e9ae7c-8cf5-404a-8403-94c5005ade7e","Type":"ContainerDied","Data":"c896ddde049b808c6d0f2bf4329657ea54a6782341b32409b2af46f0eefc7bab"} Nov 24 14:54:08 crc kubenswrapper[4822]: I1124 14:54:08.106870 4822 scope.go:117] "RemoveContainer" containerID="822a597b520e7a18214ee4487b78524bac10e3725fb3a0d2795a6cf5f81a9eaf" Nov 24 14:54:08 crc kubenswrapper[4822]: I1124 14:54:08.106893 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mzfn" Nov 24 14:54:08 crc kubenswrapper[4822]: I1124 14:54:08.132728 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rzcn5" podStartSLOduration=2.458373485 podStartE2EDuration="6.132666777s" podCreationTimestamp="2025-11-24 14:54:02 +0000 UTC" firstStartedPulling="2025-11-24 14:54:04.057263795 +0000 UTC m=+2081.173904272" lastFinishedPulling="2025-11-24 14:54:07.731557067 +0000 UTC m=+2084.848197564" observedRunningTime="2025-11-24 14:54:08.124005381 +0000 UTC m=+2085.240645858" watchObservedRunningTime="2025-11-24 14:54:08.132666777 +0000 UTC m=+2085.249307254" Nov 24 14:54:08 crc kubenswrapper[4822]: I1124 14:54:08.146034 4822 scope.go:117] "RemoveContainer" containerID="003768f6d9db7fa91bf6d2e91e54f0d8a74c109847687629c5027c3bcc35cc70" Nov 24 14:54:08 crc kubenswrapper[4822]: I1124 14:54:08.153908 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9mzfn"] Nov 24 14:54:08 crc kubenswrapper[4822]: I1124 14:54:08.163164 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9mzfn"] Nov 24 14:54:08 crc kubenswrapper[4822]: I1124 14:54:08.164518 4822 scope.go:117] "RemoveContainer" containerID="c0e417a46e8b3c6cf52e73785427c0e4b4f04cd29a2158880558de373bdf762f" Nov 24 14:54:09 crc kubenswrapper[4822]: I1124 14:54:09.716865 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" path="/var/lib/kubelet/pods/d5e9ae7c-8cf5-404a-8403-94c5005ade7e/volumes" Nov 24 14:54:12 crc kubenswrapper[4822]: I1124 14:54:12.865972 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:12 crc kubenswrapper[4822]: I1124 14:54:12.866395 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:12 crc kubenswrapper[4822]: I1124 14:54:12.945475 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:13 crc kubenswrapper[4822]: I1124 14:54:13.221991 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:14 crc kubenswrapper[4822]: I1124 14:54:14.106922 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rzcn5"] Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.178275 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rzcn5" podUID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerName="registry-server" containerID="cri-o://c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0" gracePeriod=2 Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.691870 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.865022 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-catalog-content\") pod \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.865106 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgqhr\" (UniqueName: \"kubernetes.io/projected/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-kube-api-access-tgqhr\") pod \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.865180 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-utilities\") pod \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\" (UID: \"c8c81345-38a8-4545-abeb-b6e1c2fc2c87\") " Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.866378 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-utilities" (OuterVolumeSpecName: "utilities") pod "c8c81345-38a8-4545-abeb-b6e1c2fc2c87" (UID: "c8c81345-38a8-4545-abeb-b6e1c2fc2c87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.884573 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-kube-api-access-tgqhr" (OuterVolumeSpecName: "kube-api-access-tgqhr") pod "c8c81345-38a8-4545-abeb-b6e1c2fc2c87" (UID: "c8c81345-38a8-4545-abeb-b6e1c2fc2c87"). InnerVolumeSpecName "kube-api-access-tgqhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.941352 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8c81345-38a8-4545-abeb-b6e1c2fc2c87" (UID: "c8c81345-38a8-4545-abeb-b6e1c2fc2c87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.967647 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.967677 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgqhr\" (UniqueName: \"kubernetes.io/projected/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-kube-api-access-tgqhr\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:15 crc kubenswrapper[4822]: I1124 14:54:15.967691 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8c81345-38a8-4545-abeb-b6e1c2fc2c87-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.187452 4822 generic.go:334] "Generic (PLEG): container finished" podID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerID="c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0" exitCode=0 Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.187500 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzcn5" event={"ID":"c8c81345-38a8-4545-abeb-b6e1c2fc2c87","Type":"ContainerDied","Data":"c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0"} Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.187532 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzcn5" event={"ID":"c8c81345-38a8-4545-abeb-b6e1c2fc2c87","Type":"ContainerDied","Data":"ba91836fae5d1662da9b797c3e72f93707a63c1ac11574e4e57927068e292b2d"} Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.187555 4822 scope.go:117] "RemoveContainer" containerID="c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.187665 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzcn5" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.204429 4822 scope.go:117] "RemoveContainer" containerID="e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.217289 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rzcn5"] Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.224157 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rzcn5"] Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.237652 4822 scope.go:117] "RemoveContainer" containerID="300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.281510 4822 scope.go:117] "RemoveContainer" containerID="c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0" Nov 24 14:54:16 crc kubenswrapper[4822]: E1124 14:54:16.282360 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0\": container with ID starting with c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0 not found: ID does not exist" containerID="c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.282406 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0"} err="failed to get container status \"c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0\": rpc error: code = NotFound desc = could not find container \"c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0\": container with ID starting with c0dc2e5562f3a21d1721aec7daf2d7c4a0b743e3ace8739ed67771dfa90dafb0 not found: ID does not exist" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.283420 4822 scope.go:117] "RemoveContainer" containerID="e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f" Nov 24 14:54:16 crc kubenswrapper[4822]: E1124 14:54:16.283999 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f\": container with ID starting with e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f not found: ID does not exist" containerID="e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.284033 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f"} err="failed to get container status \"e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f\": rpc error: code = NotFound desc = could not find container \"e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f\": container with ID starting with e0456c4a4459e7a7134fcd4471ff6f6a3b2f8d924017fec33a2d5e6c5aefe84f not found: ID does not exist" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.284061 4822 scope.go:117] "RemoveContainer" containerID="300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb" Nov 24 14:54:16 crc kubenswrapper[4822]: E1124 14:54:16.284354 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb\": container with ID starting with 300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb not found: ID does not exist" containerID="300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb" Nov 24 14:54:16 crc kubenswrapper[4822]: I1124 14:54:16.284383 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb"} err="failed to get container status \"300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb\": rpc error: code = NotFound desc = could not find container \"300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb\": container with ID starting with 300c2c60e7efe9129d91184950dae64fd402649733c530b41430b3c18c8df4eb not found: ID does not exist" Nov 24 14:54:17 crc kubenswrapper[4822]: I1124 14:54:17.723800 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" path="/var/lib/kubelet/pods/c8c81345-38a8-4545-abeb-b6e1c2fc2c87/volumes" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.491354 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nf62v"] Nov 24 14:54:23 crc kubenswrapper[4822]: E1124 14:54:23.492678 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerName="extract-utilities" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.492702 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerName="extract-utilities" Nov 24 14:54:23 crc kubenswrapper[4822]: E1124 14:54:23.492734 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerName="registry-server" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.492747 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerName="registry-server" Nov 24 14:54:23 crc kubenswrapper[4822]: E1124 14:54:23.492774 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerName="registry-server" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.492787 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerName="registry-server" Nov 24 14:54:23 crc kubenswrapper[4822]: E1124 14:54:23.492804 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerName="extract-content" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.492815 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerName="extract-content" Nov 24 14:54:23 crc kubenswrapper[4822]: E1124 14:54:23.492847 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerName="extract-utilities" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.492858 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerName="extract-utilities" Nov 24 14:54:23 crc kubenswrapper[4822]: E1124 14:54:23.492885 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerName="extract-content" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.492895 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerName="extract-content" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.493192 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5e9ae7c-8cf5-404a-8403-94c5005ade7e" containerName="registry-server" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.493245 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c81345-38a8-4545-abeb-b6e1c2fc2c87" containerName="registry-server" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.495897 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.499683 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf62v"] Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.549724 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-catalog-content\") pod \"redhat-marketplace-nf62v\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.549790 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-utilities\") pod \"redhat-marketplace-nf62v\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.549984 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hk2q\" (UniqueName: \"kubernetes.io/projected/aa848599-45b2-434c-9795-a71db59a117e-kube-api-access-4hk2q\") pod \"redhat-marketplace-nf62v\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.651239 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-catalog-content\") pod \"redhat-marketplace-nf62v\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.651318 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-utilities\") pod \"redhat-marketplace-nf62v\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.651552 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hk2q\" (UniqueName: \"kubernetes.io/projected/aa848599-45b2-434c-9795-a71db59a117e-kube-api-access-4hk2q\") pod \"redhat-marketplace-nf62v\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.652611 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-catalog-content\") pod \"redhat-marketplace-nf62v\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.652619 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-utilities\") pod \"redhat-marketplace-nf62v\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.671979 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hk2q\" (UniqueName: \"kubernetes.io/projected/aa848599-45b2-434c-9795-a71db59a117e-kube-api-access-4hk2q\") pod \"redhat-marketplace-nf62v\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:23 crc kubenswrapper[4822]: I1124 14:54:23.828742 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:24 crc kubenswrapper[4822]: I1124 14:54:24.408682 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf62v"] Nov 24 14:54:25 crc kubenswrapper[4822]: I1124 14:54:25.300621 4822 generic.go:334] "Generic (PLEG): container finished" podID="aa848599-45b2-434c-9795-a71db59a117e" containerID="8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175" exitCode=0 Nov 24 14:54:25 crc kubenswrapper[4822]: I1124 14:54:25.300667 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf62v" event={"ID":"aa848599-45b2-434c-9795-a71db59a117e","Type":"ContainerDied","Data":"8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175"} Nov 24 14:54:25 crc kubenswrapper[4822]: I1124 14:54:25.300920 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf62v" event={"ID":"aa848599-45b2-434c-9795-a71db59a117e","Type":"ContainerStarted","Data":"cb2397a44a8d91d0865a8a5383ba6a2332d34d94e519f6f65816314339deb857"} Nov 24 14:54:26 crc kubenswrapper[4822]: I1124 14:54:26.314613 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf62v" event={"ID":"aa848599-45b2-434c-9795-a71db59a117e","Type":"ContainerStarted","Data":"39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3"} Nov 24 14:54:27 crc kubenswrapper[4822]: I1124 14:54:27.328657 4822 generic.go:334] "Generic (PLEG): container finished" podID="aa848599-45b2-434c-9795-a71db59a117e" containerID="39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3" exitCode=0 Nov 24 14:54:27 crc kubenswrapper[4822]: I1124 14:54:27.329036 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf62v" event={"ID":"aa848599-45b2-434c-9795-a71db59a117e","Type":"ContainerDied","Data":"39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3"} Nov 24 14:54:28 crc kubenswrapper[4822]: I1124 14:54:28.344059 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf62v" event={"ID":"aa848599-45b2-434c-9795-a71db59a117e","Type":"ContainerStarted","Data":"d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce"} Nov 24 14:54:28 crc kubenswrapper[4822]: I1124 14:54:28.371861 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nf62v" podStartSLOduration=2.897648992 podStartE2EDuration="5.371844511s" podCreationTimestamp="2025-11-24 14:54:23 +0000 UTC" firstStartedPulling="2025-11-24 14:54:25.302608029 +0000 UTC m=+2102.419248506" lastFinishedPulling="2025-11-24 14:54:27.776803508 +0000 UTC m=+2104.893444025" observedRunningTime="2025-11-24 14:54:28.363036111 +0000 UTC m=+2105.479676608" watchObservedRunningTime="2025-11-24 14:54:28.371844511 +0000 UTC m=+2105.488484988" Nov 24 14:54:33 crc kubenswrapper[4822]: I1124 14:54:33.829936 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:33 crc kubenswrapper[4822]: I1124 14:54:33.830524 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:33 crc kubenswrapper[4822]: I1124 14:54:33.890705 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:34 crc kubenswrapper[4822]: I1124 14:54:34.474145 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:34 crc kubenswrapper[4822]: I1124 14:54:34.530315 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf62v"] Nov 24 14:54:36 crc kubenswrapper[4822]: I1124 14:54:36.433854 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nf62v" podUID="aa848599-45b2-434c-9795-a71db59a117e" containerName="registry-server" containerID="cri-o://d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce" gracePeriod=2 Nov 24 14:54:36 crc kubenswrapper[4822]: I1124 14:54:36.926222 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:36 crc kubenswrapper[4822]: I1124 14:54:36.971956 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hk2q\" (UniqueName: \"kubernetes.io/projected/aa848599-45b2-434c-9795-a71db59a117e-kube-api-access-4hk2q\") pod \"aa848599-45b2-434c-9795-a71db59a117e\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " Nov 24 14:54:36 crc kubenswrapper[4822]: I1124 14:54:36.972063 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-utilities\") pod \"aa848599-45b2-434c-9795-a71db59a117e\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " Nov 24 14:54:36 crc kubenswrapper[4822]: I1124 14:54:36.972244 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-catalog-content\") pod \"aa848599-45b2-434c-9795-a71db59a117e\" (UID: \"aa848599-45b2-434c-9795-a71db59a117e\") " Nov 24 14:54:36 crc kubenswrapper[4822]: I1124 14:54:36.973682 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-utilities" (OuterVolumeSpecName: "utilities") pod "aa848599-45b2-434c-9795-a71db59a117e" (UID: "aa848599-45b2-434c-9795-a71db59a117e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:36 crc kubenswrapper[4822]: I1124 14:54:36.981039 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa848599-45b2-434c-9795-a71db59a117e-kube-api-access-4hk2q" (OuterVolumeSpecName: "kube-api-access-4hk2q") pod "aa848599-45b2-434c-9795-a71db59a117e" (UID: "aa848599-45b2-434c-9795-a71db59a117e"). InnerVolumeSpecName "kube-api-access-4hk2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.001505 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa848599-45b2-434c-9795-a71db59a117e" (UID: "aa848599-45b2-434c-9795-a71db59a117e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.074233 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.074260 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hk2q\" (UniqueName: \"kubernetes.io/projected/aa848599-45b2-434c-9795-a71db59a117e-kube-api-access-4hk2q\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.074271 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa848599-45b2-434c-9795-a71db59a117e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.449937 4822 generic.go:334] "Generic (PLEG): container finished" podID="aa848599-45b2-434c-9795-a71db59a117e" containerID="d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce" exitCode=0 Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.450015 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf62v" event={"ID":"aa848599-45b2-434c-9795-a71db59a117e","Type":"ContainerDied","Data":"d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce"} Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.450065 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf62v" event={"ID":"aa848599-45b2-434c-9795-a71db59a117e","Type":"ContainerDied","Data":"cb2397a44a8d91d0865a8a5383ba6a2332d34d94e519f6f65816314339deb857"} Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.450094 4822 scope.go:117] "RemoveContainer" containerID="d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.450098 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf62v" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.473569 4822 scope.go:117] "RemoveContainer" containerID="39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.502917 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf62v"] Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.512021 4822 scope.go:117] "RemoveContainer" containerID="8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.515596 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf62v"] Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.573103 4822 scope.go:117] "RemoveContainer" containerID="d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce" Nov 24 14:54:37 crc kubenswrapper[4822]: E1124 14:54:37.573550 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce\": container with ID starting with d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce not found: ID does not exist" containerID="d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.573582 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce"} err="failed to get container status \"d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce\": rpc error: code = NotFound desc = could not find container \"d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce\": container with ID starting with d54c11ff1e4f792efd7fda7aa5221034fbc47bfcb88eb69a5b703060099340ce not found: ID does not exist" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.573603 4822 scope.go:117] "RemoveContainer" containerID="39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3" Nov 24 14:54:37 crc kubenswrapper[4822]: E1124 14:54:37.573860 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3\": container with ID starting with 39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3 not found: ID does not exist" containerID="39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.573918 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3"} err="failed to get container status \"39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3\": rpc error: code = NotFound desc = could not find container \"39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3\": container with ID starting with 39d5fb9cd02827319015c1ac476025e6556f9694454de77e3837090701b32fc3 not found: ID does not exist" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.573964 4822 scope.go:117] "RemoveContainer" containerID="8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175" Nov 24 14:54:37 crc kubenswrapper[4822]: E1124 14:54:37.574524 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175\": container with ID starting with 8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175 not found: ID does not exist" containerID="8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.574551 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175"} err="failed to get container status \"8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175\": rpc error: code = NotFound desc = could not find container \"8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175\": container with ID starting with 8a3db224d022ab251e6969e08a56c89cf4e5df7e7f805057700f113cc15a6175 not found: ID does not exist" Nov 24 14:54:37 crc kubenswrapper[4822]: I1124 14:54:37.722333 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa848599-45b2-434c-9795-a71db59a117e" path="/var/lib/kubelet/pods/aa848599-45b2-434c-9795-a71db59a117e/volumes" Nov 24 14:55:41 crc kubenswrapper[4822]: I1124 14:55:41.578510 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:55:41 crc kubenswrapper[4822]: I1124 14:55:41.579044 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:55:49 crc kubenswrapper[4822]: I1124 14:55:49.337836 4822 generic.go:334] "Generic (PLEG): container finished" podID="3cea7046-1628-413a-99d0-aa36783ec7fc" containerID="05e1ca3e885f17ba77ab4700f973ff9b8fbe9382da72903ede36ca92f7ce2964" exitCode=0 Nov 24 14:55:49 crc kubenswrapper[4822]: I1124 14:55:49.337908 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" event={"ID":"3cea7046-1628-413a-99d0-aa36783ec7fc","Type":"ContainerDied","Data":"05e1ca3e885f17ba77ab4700f973ff9b8fbe9382da72903ede36ca92f7ce2964"} Nov 24 14:55:50 crc kubenswrapper[4822]: I1124 14:55:50.917752 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.081473 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lwzn\" (UniqueName: \"kubernetes.io/projected/3cea7046-1628-413a-99d0-aa36783ec7fc-kube-api-access-6lwzn\") pod \"3cea7046-1628-413a-99d0-aa36783ec7fc\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.084041 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-bootstrap-combined-ca-bundle\") pod \"3cea7046-1628-413a-99d0-aa36783ec7fc\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.085007 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-inventory\") pod \"3cea7046-1628-413a-99d0-aa36783ec7fc\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.085427 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-ssh-key\") pod \"3cea7046-1628-413a-99d0-aa36783ec7fc\" (UID: \"3cea7046-1628-413a-99d0-aa36783ec7fc\") " Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.088024 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3cea7046-1628-413a-99d0-aa36783ec7fc" (UID: "3cea7046-1628-413a-99d0-aa36783ec7fc"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.089772 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cea7046-1628-413a-99d0-aa36783ec7fc-kube-api-access-6lwzn" (OuterVolumeSpecName: "kube-api-access-6lwzn") pod "3cea7046-1628-413a-99d0-aa36783ec7fc" (UID: "3cea7046-1628-413a-99d0-aa36783ec7fc"). InnerVolumeSpecName "kube-api-access-6lwzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.113833 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-inventory" (OuterVolumeSpecName: "inventory") pod "3cea7046-1628-413a-99d0-aa36783ec7fc" (UID: "3cea7046-1628-413a-99d0-aa36783ec7fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.119789 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3cea7046-1628-413a-99d0-aa36783ec7fc" (UID: "3cea7046-1628-413a-99d0-aa36783ec7fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.188820 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lwzn\" (UniqueName: \"kubernetes.io/projected/3cea7046-1628-413a-99d0-aa36783ec7fc-kube-api-access-6lwzn\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.188848 4822 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.188857 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.188866 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cea7046-1628-413a-99d0-aa36783ec7fc-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.367377 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" event={"ID":"3cea7046-1628-413a-99d0-aa36783ec7fc","Type":"ContainerDied","Data":"b9083e946b1fcb6f85c91a0711bf2e9f6ac9a9abe16e8067f0e9a9d6ab53ae06"} Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.367474 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9083e946b1fcb6f85c91a0711bf2e9f6ac9a9abe16e8067f0e9a9d6ab53ae06" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.367532 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.488857 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5"] Nov 24 14:55:51 crc kubenswrapper[4822]: E1124 14:55:51.489861 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cea7046-1628-413a-99d0-aa36783ec7fc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.489909 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cea7046-1628-413a-99d0-aa36783ec7fc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 14:55:51 crc kubenswrapper[4822]: E1124 14:55:51.489978 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa848599-45b2-434c-9795-a71db59a117e" containerName="extract-utilities" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.489993 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa848599-45b2-434c-9795-a71db59a117e" containerName="extract-utilities" Nov 24 14:55:51 crc kubenswrapper[4822]: E1124 14:55:51.490014 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa848599-45b2-434c-9795-a71db59a117e" containerName="registry-server" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.490027 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa848599-45b2-434c-9795-a71db59a117e" containerName="registry-server" Nov 24 14:55:51 crc kubenswrapper[4822]: E1124 14:55:51.490052 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa848599-45b2-434c-9795-a71db59a117e" containerName="extract-content" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.490065 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa848599-45b2-434c-9795-a71db59a117e" containerName="extract-content" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.490476 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa848599-45b2-434c-9795-a71db59a117e" containerName="registry-server" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.490522 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cea7046-1628-413a-99d0-aa36783ec7fc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.492038 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.495230 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.495538 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.497762 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.497935 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.509500 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5"] Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.601642 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.601765 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.601821 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp9vf\" (UniqueName: \"kubernetes.io/projected/9700df2b-526c-44d1-ba51-f9313739f6a8-kube-api-access-wp9vf\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.703928 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.704046 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.704101 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp9vf\" (UniqueName: \"kubernetes.io/projected/9700df2b-526c-44d1-ba51-f9313739f6a8-kube-api-access-wp9vf\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.707898 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.708173 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.731841 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp9vf\" (UniqueName: \"kubernetes.io/projected/9700df2b-526c-44d1-ba51-f9313739f6a8-kube-api-access-wp9vf\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:51 crc kubenswrapper[4822]: I1124 14:55:51.836509 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:55:52 crc kubenswrapper[4822]: W1124 14:55:52.452043 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9700df2b_526c_44d1_ba51_f9313739f6a8.slice/crio-ee461a26be8f5724581897e79a1f71593cc85c5333c54c8e8896ca575d7f5a0d WatchSource:0}: Error finding container ee461a26be8f5724581897e79a1f71593cc85c5333c54c8e8896ca575d7f5a0d: Status 404 returned error can't find the container with id ee461a26be8f5724581897e79a1f71593cc85c5333c54c8e8896ca575d7f5a0d Nov 24 14:55:52 crc kubenswrapper[4822]: I1124 14:55:52.463182 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5"] Nov 24 14:55:53 crc kubenswrapper[4822]: I1124 14:55:53.400608 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" event={"ID":"9700df2b-526c-44d1-ba51-f9313739f6a8","Type":"ContainerStarted","Data":"3b1ad52609e388d80fdca77db8ec88a9eda0f6ec48a82060e6d3d69f4a73ae4f"} Nov 24 14:55:53 crc kubenswrapper[4822]: I1124 14:55:53.401227 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" event={"ID":"9700df2b-526c-44d1-ba51-f9313739f6a8","Type":"ContainerStarted","Data":"ee461a26be8f5724581897e79a1f71593cc85c5333c54c8e8896ca575d7f5a0d"} Nov 24 14:55:53 crc kubenswrapper[4822]: I1124 14:55:53.429630 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" podStartSLOduration=1.93367831 podStartE2EDuration="2.429604958s" podCreationTimestamp="2025-11-24 14:55:51 +0000 UTC" firstStartedPulling="2025-11-24 14:55:52.457868132 +0000 UTC m=+2189.574508619" lastFinishedPulling="2025-11-24 14:55:52.95379479 +0000 UTC m=+2190.070435267" observedRunningTime="2025-11-24 14:55:53.420168567 +0000 UTC m=+2190.536809114" watchObservedRunningTime="2025-11-24 14:55:53.429604958 +0000 UTC m=+2190.546245475" Nov 24 14:56:11 crc kubenswrapper[4822]: I1124 14:56:11.578112 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:56:11 crc kubenswrapper[4822]: I1124 14:56:11.578717 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.372629 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nqfc7"] Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.376864 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.391590 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nqfc7"] Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.507020 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhpd8\" (UniqueName: \"kubernetes.io/projected/c851ebfd-7217-4a05-af0f-c25923a311af-kube-api-access-jhpd8\") pod \"redhat-operators-nqfc7\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.507152 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-catalog-content\") pod \"redhat-operators-nqfc7\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.507224 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-utilities\") pod \"redhat-operators-nqfc7\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.609007 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-catalog-content\") pod \"redhat-operators-nqfc7\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.609391 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-utilities\") pod \"redhat-operators-nqfc7\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.609486 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhpd8\" (UniqueName: \"kubernetes.io/projected/c851ebfd-7217-4a05-af0f-c25923a311af-kube-api-access-jhpd8\") pod \"redhat-operators-nqfc7\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.609770 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-catalog-content\") pod \"redhat-operators-nqfc7\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.610064 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-utilities\") pod \"redhat-operators-nqfc7\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.630616 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhpd8\" (UniqueName: \"kubernetes.io/projected/c851ebfd-7217-4a05-af0f-c25923a311af-kube-api-access-jhpd8\") pod \"redhat-operators-nqfc7\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:24 crc kubenswrapper[4822]: I1124 14:56:24.781254 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:25 crc kubenswrapper[4822]: I1124 14:56:25.220166 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nqfc7"] Nov 24 14:56:25 crc kubenswrapper[4822]: W1124 14:56:25.221895 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc851ebfd_7217_4a05_af0f_c25923a311af.slice/crio-1ddee8b101f8a2091b9869ad82ceb0395093962ee382b4b94d1dbc3052b8518d WatchSource:0}: Error finding container 1ddee8b101f8a2091b9869ad82ceb0395093962ee382b4b94d1dbc3052b8518d: Status 404 returned error can't find the container with id 1ddee8b101f8a2091b9869ad82ceb0395093962ee382b4b94d1dbc3052b8518d Nov 24 14:56:25 crc kubenswrapper[4822]: I1124 14:56:25.758043 4822 generic.go:334] "Generic (PLEG): container finished" podID="c851ebfd-7217-4a05-af0f-c25923a311af" containerID="ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1" exitCode=0 Nov 24 14:56:25 crc kubenswrapper[4822]: I1124 14:56:25.758105 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqfc7" event={"ID":"c851ebfd-7217-4a05-af0f-c25923a311af","Type":"ContainerDied","Data":"ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1"} Nov 24 14:56:25 crc kubenswrapper[4822]: I1124 14:56:25.758383 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqfc7" event={"ID":"c851ebfd-7217-4a05-af0f-c25923a311af","Type":"ContainerStarted","Data":"1ddee8b101f8a2091b9869ad82ceb0395093962ee382b4b94d1dbc3052b8518d"} Nov 24 14:56:27 crc kubenswrapper[4822]: I1124 14:56:27.789065 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqfc7" event={"ID":"c851ebfd-7217-4a05-af0f-c25923a311af","Type":"ContainerStarted","Data":"3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492"} Nov 24 14:56:30 crc kubenswrapper[4822]: I1124 14:56:30.833878 4822 generic.go:334] "Generic (PLEG): container finished" podID="c851ebfd-7217-4a05-af0f-c25923a311af" containerID="3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492" exitCode=0 Nov 24 14:56:30 crc kubenswrapper[4822]: I1124 14:56:30.833920 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqfc7" event={"ID":"c851ebfd-7217-4a05-af0f-c25923a311af","Type":"ContainerDied","Data":"3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492"} Nov 24 14:56:31 crc kubenswrapper[4822]: I1124 14:56:31.847973 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqfc7" event={"ID":"c851ebfd-7217-4a05-af0f-c25923a311af","Type":"ContainerStarted","Data":"c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c"} Nov 24 14:56:31 crc kubenswrapper[4822]: I1124 14:56:31.879321 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nqfc7" podStartSLOduration=2.355982789 podStartE2EDuration="7.879301639s" podCreationTimestamp="2025-11-24 14:56:24 +0000 UTC" firstStartedPulling="2025-11-24 14:56:25.759713081 +0000 UTC m=+2222.876353558" lastFinishedPulling="2025-11-24 14:56:31.283031931 +0000 UTC m=+2228.399672408" observedRunningTime="2025-11-24 14:56:31.868938149 +0000 UTC m=+2228.985578636" watchObservedRunningTime="2025-11-24 14:56:31.879301639 +0000 UTC m=+2228.995942116" Nov 24 14:56:34 crc kubenswrapper[4822]: I1124 14:56:34.782447 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:34 crc kubenswrapper[4822]: I1124 14:56:34.782993 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:35 crc kubenswrapper[4822]: I1124 14:56:35.846666 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nqfc7" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" containerName="registry-server" probeResult="failure" output=< Nov 24 14:56:35 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 14:56:35 crc kubenswrapper[4822]: > Nov 24 14:56:41 crc kubenswrapper[4822]: I1124 14:56:41.578331 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:56:41 crc kubenswrapper[4822]: I1124 14:56:41.578768 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:56:41 crc kubenswrapper[4822]: I1124 14:56:41.578852 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 14:56:41 crc kubenswrapper[4822]: I1124 14:56:41.580007 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:56:41 crc kubenswrapper[4822]: I1124 14:56:41.580094 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" gracePeriod=600 Nov 24 14:56:41 crc kubenswrapper[4822]: E1124 14:56:41.710518 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:56:41 crc kubenswrapper[4822]: I1124 14:56:41.958292 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" exitCode=0 Nov 24 14:56:41 crc kubenswrapper[4822]: I1124 14:56:41.958353 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300"} Nov 24 14:56:41 crc kubenswrapper[4822]: I1124 14:56:41.958399 4822 scope.go:117] "RemoveContainer" containerID="5cc0b1343a16a8d188981e3c048d5b5d39678fa534e29571730f78b864e29cf8" Nov 24 14:56:41 crc kubenswrapper[4822]: I1124 14:56:41.959628 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:56:41 crc kubenswrapper[4822]: E1124 14:56:41.959924 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:56:44 crc kubenswrapper[4822]: I1124 14:56:44.912766 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:44 crc kubenswrapper[4822]: I1124 14:56:44.971448 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:45.999915 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nqfc7"] Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.006620 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nqfc7" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" containerName="registry-server" containerID="cri-o://c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c" gracePeriod=2 Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.578860 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.770014 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-utilities\") pod \"c851ebfd-7217-4a05-af0f-c25923a311af\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.770238 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-catalog-content\") pod \"c851ebfd-7217-4a05-af0f-c25923a311af\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.771362 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhpd8\" (UniqueName: \"kubernetes.io/projected/c851ebfd-7217-4a05-af0f-c25923a311af-kube-api-access-jhpd8\") pod \"c851ebfd-7217-4a05-af0f-c25923a311af\" (UID: \"c851ebfd-7217-4a05-af0f-c25923a311af\") " Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.771759 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-utilities" (OuterVolumeSpecName: "utilities") pod "c851ebfd-7217-4a05-af0f-c25923a311af" (UID: "c851ebfd-7217-4a05-af0f-c25923a311af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.772650 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.780401 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c851ebfd-7217-4a05-af0f-c25923a311af-kube-api-access-jhpd8" (OuterVolumeSpecName: "kube-api-access-jhpd8") pod "c851ebfd-7217-4a05-af0f-c25923a311af" (UID: "c851ebfd-7217-4a05-af0f-c25923a311af"). InnerVolumeSpecName "kube-api-access-jhpd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.858116 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c851ebfd-7217-4a05-af0f-c25923a311af" (UID: "c851ebfd-7217-4a05-af0f-c25923a311af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.875536 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c851ebfd-7217-4a05-af0f-c25923a311af-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:46 crc kubenswrapper[4822]: I1124 14:56:46.875594 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhpd8\" (UniqueName: \"kubernetes.io/projected/c851ebfd-7217-4a05-af0f-c25923a311af-kube-api-access-jhpd8\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.017285 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqfc7" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.017295 4822 generic.go:334] "Generic (PLEG): container finished" podID="c851ebfd-7217-4a05-af0f-c25923a311af" containerID="c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c" exitCode=0 Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.017343 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqfc7" event={"ID":"c851ebfd-7217-4a05-af0f-c25923a311af","Type":"ContainerDied","Data":"c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c"} Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.017381 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqfc7" event={"ID":"c851ebfd-7217-4a05-af0f-c25923a311af","Type":"ContainerDied","Data":"1ddee8b101f8a2091b9869ad82ceb0395093962ee382b4b94d1dbc3052b8518d"} Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.017403 4822 scope.go:117] "RemoveContainer" containerID="c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.045891 4822 scope.go:117] "RemoveContainer" containerID="3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.060024 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nqfc7"] Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.071630 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nqfc7"] Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.078231 4822 scope.go:117] "RemoveContainer" containerID="ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.127144 4822 scope.go:117] "RemoveContainer" containerID="c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c" Nov 24 14:56:47 crc kubenswrapper[4822]: E1124 14:56:47.127552 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c\": container with ID starting with c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c not found: ID does not exist" containerID="c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.127585 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c"} err="failed to get container status \"c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c\": rpc error: code = NotFound desc = could not find container \"c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c\": container with ID starting with c570f05c28d9ee667fe0da2808bd62994b3d880b76fce46a9b070ce3ca24ee0c not found: ID does not exist" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.127608 4822 scope.go:117] "RemoveContainer" containerID="3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492" Nov 24 14:56:47 crc kubenswrapper[4822]: E1124 14:56:47.127792 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492\": container with ID starting with 3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492 not found: ID does not exist" containerID="3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.127814 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492"} err="failed to get container status \"3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492\": rpc error: code = NotFound desc = could not find container \"3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492\": container with ID starting with 3ed607a4a9bf6f70e3c00928a9b5c959431767fcf2a67e32600b58479123a492 not found: ID does not exist" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.127827 4822 scope.go:117] "RemoveContainer" containerID="ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1" Nov 24 14:56:47 crc kubenswrapper[4822]: E1124 14:56:47.127987 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1\": container with ID starting with ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1 not found: ID does not exist" containerID="ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.128003 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1"} err="failed to get container status \"ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1\": rpc error: code = NotFound desc = could not find container \"ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1\": container with ID starting with ef8d255a37423077713f15abde2e6f7f2734cfd9b5694209f1a2efc4d31be4c1 not found: ID does not exist" Nov 24 14:56:47 crc kubenswrapper[4822]: I1124 14:56:47.720921 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" path="/var/lib/kubelet/pods/c851ebfd-7217-4a05-af0f-c25923a311af/volumes" Nov 24 14:56:52 crc kubenswrapper[4822]: I1124 14:56:52.704912 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:56:52 crc kubenswrapper[4822]: E1124 14:56:52.705826 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:57:04 crc kubenswrapper[4822]: I1124 14:57:04.704544 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:57:04 crc kubenswrapper[4822]: E1124 14:57:04.705385 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:57:16 crc kubenswrapper[4822]: I1124 14:57:16.705235 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:57:16 crc kubenswrapper[4822]: E1124 14:57:16.706034 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:57:27 crc kubenswrapper[4822]: I1124 14:57:27.705020 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:57:27 crc kubenswrapper[4822]: E1124 14:57:27.705969 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:57:38 crc kubenswrapper[4822]: I1124 14:57:38.705798 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:57:38 crc kubenswrapper[4822]: E1124 14:57:38.707106 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:57:49 crc kubenswrapper[4822]: I1124 14:57:49.691121 4822 generic.go:334] "Generic (PLEG): container finished" podID="9700df2b-526c-44d1-ba51-f9313739f6a8" containerID="3b1ad52609e388d80fdca77db8ec88a9eda0f6ec48a82060e6d3d69f4a73ae4f" exitCode=0 Nov 24 14:57:49 crc kubenswrapper[4822]: I1124 14:57:49.691224 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" event={"ID":"9700df2b-526c-44d1-ba51-f9313739f6a8","Type":"ContainerDied","Data":"3b1ad52609e388d80fdca77db8ec88a9eda0f6ec48a82060e6d3d69f4a73ae4f"} Nov 24 14:57:49 crc kubenswrapper[4822]: I1124 14:57:49.706118 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:57:49 crc kubenswrapper[4822]: E1124 14:57:49.706588 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.209228 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.354016 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-ssh-key\") pod \"9700df2b-526c-44d1-ba51-f9313739f6a8\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.354125 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp9vf\" (UniqueName: \"kubernetes.io/projected/9700df2b-526c-44d1-ba51-f9313739f6a8-kube-api-access-wp9vf\") pod \"9700df2b-526c-44d1-ba51-f9313739f6a8\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.354428 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-inventory\") pod \"9700df2b-526c-44d1-ba51-f9313739f6a8\" (UID: \"9700df2b-526c-44d1-ba51-f9313739f6a8\") " Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.361422 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9700df2b-526c-44d1-ba51-f9313739f6a8-kube-api-access-wp9vf" (OuterVolumeSpecName: "kube-api-access-wp9vf") pod "9700df2b-526c-44d1-ba51-f9313739f6a8" (UID: "9700df2b-526c-44d1-ba51-f9313739f6a8"). InnerVolumeSpecName "kube-api-access-wp9vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.400684 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9700df2b-526c-44d1-ba51-f9313739f6a8" (UID: "9700df2b-526c-44d1-ba51-f9313739f6a8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.400879 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-inventory" (OuterVolumeSpecName: "inventory") pod "9700df2b-526c-44d1-ba51-f9313739f6a8" (UID: "9700df2b-526c-44d1-ba51-f9313739f6a8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.458041 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.458078 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9700df2b-526c-44d1-ba51-f9313739f6a8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.458093 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp9vf\" (UniqueName: \"kubernetes.io/projected/9700df2b-526c-44d1-ba51-f9313739f6a8-kube-api-access-wp9vf\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.714685 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.730534 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5" event={"ID":"9700df2b-526c-44d1-ba51-f9313739f6a8","Type":"ContainerDied","Data":"ee461a26be8f5724581897e79a1f71593cc85c5333c54c8e8896ca575d7f5a0d"} Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.732538 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee461a26be8f5724581897e79a1f71593cc85c5333c54c8e8896ca575d7f5a0d" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.831607 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj"] Nov 24 14:57:51 crc kubenswrapper[4822]: E1124 14:57:51.832057 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9700df2b-526c-44d1-ba51-f9313739f6a8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.832077 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9700df2b-526c-44d1-ba51-f9313739f6a8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 14:57:51 crc kubenswrapper[4822]: E1124 14:57:51.832103 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" containerName="extract-utilities" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.832111 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" containerName="extract-utilities" Nov 24 14:57:51 crc kubenswrapper[4822]: E1124 14:57:51.832123 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" containerName="extract-content" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.832129 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" containerName="extract-content" Nov 24 14:57:51 crc kubenswrapper[4822]: E1124 14:57:51.832141 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" containerName="registry-server" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.832147 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" containerName="registry-server" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.832340 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="9700df2b-526c-44d1-ba51-f9313739f6a8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.832372 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="c851ebfd-7217-4a05-af0f-c25923a311af" containerName="registry-server" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.833080 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.836037 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.836273 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.836840 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.837215 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.842795 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj"] Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.979858 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x8khj\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.979936 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x8khj\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:51 crc kubenswrapper[4822]: I1124 14:57:51.980022 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2g2f\" (UniqueName: \"kubernetes.io/projected/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-kube-api-access-b2g2f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x8khj\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:52 crc kubenswrapper[4822]: I1124 14:57:52.081605 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x8khj\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:52 crc kubenswrapper[4822]: I1124 14:57:52.081720 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2g2f\" (UniqueName: \"kubernetes.io/projected/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-kube-api-access-b2g2f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x8khj\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:52 crc kubenswrapper[4822]: I1124 14:57:52.081831 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x8khj\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:52 crc kubenswrapper[4822]: I1124 14:57:52.088435 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x8khj\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:52 crc kubenswrapper[4822]: I1124 14:57:52.090699 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x8khj\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:52 crc kubenswrapper[4822]: I1124 14:57:52.105620 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2g2f\" (UniqueName: \"kubernetes.io/projected/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-kube-api-access-b2g2f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x8khj\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:52 crc kubenswrapper[4822]: I1124 14:57:52.152253 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:57:52 crc kubenswrapper[4822]: I1124 14:57:52.772536 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj"] Nov 24 14:57:52 crc kubenswrapper[4822]: I1124 14:57:52.774016 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:57:53 crc kubenswrapper[4822]: I1124 14:57:53.734514 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" event={"ID":"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f","Type":"ContainerStarted","Data":"46abdab291eef00a343033bb012924e5d234cb6a6f75ce4ce648a49069bb4bff"} Nov 24 14:57:54 crc kubenswrapper[4822]: I1124 14:57:54.765753 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" event={"ID":"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f","Type":"ContainerStarted","Data":"fc425ceb974c074e3030ab4e59293eacc85e681994a3af9028d3d5c31609f571"} Nov 24 14:57:54 crc kubenswrapper[4822]: I1124 14:57:54.816557 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" podStartSLOduration=3.086930949 podStartE2EDuration="3.816536355s" podCreationTimestamp="2025-11-24 14:57:51 +0000 UTC" firstStartedPulling="2025-11-24 14:57:52.773625157 +0000 UTC m=+2309.890265644" lastFinishedPulling="2025-11-24 14:57:53.503230563 +0000 UTC m=+2310.619871050" observedRunningTime="2025-11-24 14:57:54.797021598 +0000 UTC m=+2311.913662145" watchObservedRunningTime="2025-11-24 14:57:54.816536355 +0000 UTC m=+2311.933176842" Nov 24 14:58:02 crc kubenswrapper[4822]: I1124 14:58:02.705291 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:58:02 crc kubenswrapper[4822]: E1124 14:58:02.706030 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:58:17 crc kubenswrapper[4822]: I1124 14:58:17.704137 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:58:17 crc kubenswrapper[4822]: E1124 14:58:17.704958 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:58:28 crc kubenswrapper[4822]: I1124 14:58:28.704996 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:58:28 crc kubenswrapper[4822]: E1124 14:58:28.706030 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:58:42 crc kubenswrapper[4822]: I1124 14:58:42.704762 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:58:42 crc kubenswrapper[4822]: E1124 14:58:42.705862 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:58:56 crc kubenswrapper[4822]: I1124 14:58:56.705284 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:58:56 crc kubenswrapper[4822]: E1124 14:58:56.706360 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:59:09 crc kubenswrapper[4822]: I1124 14:59:09.705074 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:59:09 crc kubenswrapper[4822]: E1124 14:59:09.705835 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:59:12 crc kubenswrapper[4822]: I1124 14:59:12.592308 4822 generic.go:334] "Generic (PLEG): container finished" podID="dfb4aa15-44b2-45f7-b4a0-2b850bffa95f" containerID="fc425ceb974c074e3030ab4e59293eacc85e681994a3af9028d3d5c31609f571" exitCode=0 Nov 24 14:59:12 crc kubenswrapper[4822]: I1124 14:59:12.592418 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" event={"ID":"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f","Type":"ContainerDied","Data":"fc425ceb974c074e3030ab4e59293eacc85e681994a3af9028d3d5c31609f571"} Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.071389 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.220010 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-inventory\") pod \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.220147 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-ssh-key\") pod \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.220294 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2g2f\" (UniqueName: \"kubernetes.io/projected/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-kube-api-access-b2g2f\") pod \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\" (UID: \"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f\") " Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.235601 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-kube-api-access-b2g2f" (OuterVolumeSpecName: "kube-api-access-b2g2f") pod "dfb4aa15-44b2-45f7-b4a0-2b850bffa95f" (UID: "dfb4aa15-44b2-45f7-b4a0-2b850bffa95f"). InnerVolumeSpecName "kube-api-access-b2g2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.278479 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-inventory" (OuterVolumeSpecName: "inventory") pod "dfb4aa15-44b2-45f7-b4a0-2b850bffa95f" (UID: "dfb4aa15-44b2-45f7-b4a0-2b850bffa95f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.287484 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dfb4aa15-44b2-45f7-b4a0-2b850bffa95f" (UID: "dfb4aa15-44b2-45f7-b4a0-2b850bffa95f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.322739 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2g2f\" (UniqueName: \"kubernetes.io/projected/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-kube-api-access-b2g2f\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.322780 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.322794 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb4aa15-44b2-45f7-b4a0-2b850bffa95f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.614630 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" event={"ID":"dfb4aa15-44b2-45f7-b4a0-2b850bffa95f","Type":"ContainerDied","Data":"46abdab291eef00a343033bb012924e5d234cb6a6f75ce4ce648a49069bb4bff"} Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.614670 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x8khj" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.614674 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46abdab291eef00a343033bb012924e5d234cb6a6f75ce4ce648a49069bb4bff" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.706118 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k"] Nov 24 14:59:14 crc kubenswrapper[4822]: E1124 14:59:14.706586 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb4aa15-44b2-45f7-b4a0-2b850bffa95f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.706604 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb4aa15-44b2-45f7-b4a0-2b850bffa95f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.706803 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfb4aa15-44b2-45f7-b4a0-2b850bffa95f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.707552 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.710652 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.712784 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.712817 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.715927 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.720313 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k"] Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.833129 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.833187 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqnnf\" (UniqueName: \"kubernetes.io/projected/3b528ac3-bde2-404d-b02b-9e7a56710af4-kube-api-access-fqnnf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.833443 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.935869 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.936937 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.936994 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqnnf\" (UniqueName: \"kubernetes.io/projected/3b528ac3-bde2-404d-b02b-9e7a56710af4-kube-api-access-fqnnf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.939944 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.941462 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:14 crc kubenswrapper[4822]: I1124 14:59:14.959748 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqnnf\" (UniqueName: \"kubernetes.io/projected/3b528ac3-bde2-404d-b02b-9e7a56710af4-kube-api-access-fqnnf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:15 crc kubenswrapper[4822]: I1124 14:59:15.032892 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:15 crc kubenswrapper[4822]: I1124 14:59:15.595260 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k"] Nov 24 14:59:15 crc kubenswrapper[4822]: I1124 14:59:15.623760 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" event={"ID":"3b528ac3-bde2-404d-b02b-9e7a56710af4","Type":"ContainerStarted","Data":"18cc7c33e36eb51f3f41b7ddaf6451c9979588a555f94d72fe3575f00bd4203c"} Nov 24 14:59:17 crc kubenswrapper[4822]: I1124 14:59:17.660069 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" event={"ID":"3b528ac3-bde2-404d-b02b-9e7a56710af4","Type":"ContainerStarted","Data":"43ffe9ff07b01cdf3d3be4591fc7a66458ec6580d5f073c7d7f5f03adf74dfd6"} Nov 24 14:59:17 crc kubenswrapper[4822]: I1124 14:59:17.686357 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" podStartSLOduration=2.703673459 podStartE2EDuration="3.686335665s" podCreationTimestamp="2025-11-24 14:59:14 +0000 UTC" firstStartedPulling="2025-11-24 14:59:15.603632952 +0000 UTC m=+2392.720273429" lastFinishedPulling="2025-11-24 14:59:16.586295128 +0000 UTC m=+2393.702935635" observedRunningTime="2025-11-24 14:59:17.677549361 +0000 UTC m=+2394.794189858" watchObservedRunningTime="2025-11-24 14:59:17.686335665 +0000 UTC m=+2394.802976152" Nov 24 14:59:22 crc kubenswrapper[4822]: I1124 14:59:22.704983 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:59:22 crc kubenswrapper[4822]: E1124 14:59:22.705867 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:59:22 crc kubenswrapper[4822]: I1124 14:59:22.712913 4822 generic.go:334] "Generic (PLEG): container finished" podID="3b528ac3-bde2-404d-b02b-9e7a56710af4" containerID="43ffe9ff07b01cdf3d3be4591fc7a66458ec6580d5f073c7d7f5f03adf74dfd6" exitCode=0 Nov 24 14:59:22 crc kubenswrapper[4822]: I1124 14:59:22.712965 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" event={"ID":"3b528ac3-bde2-404d-b02b-9e7a56710af4","Type":"ContainerDied","Data":"43ffe9ff07b01cdf3d3be4591fc7a66458ec6580d5f073c7d7f5f03adf74dfd6"} Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.178229 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.275469 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqnnf\" (UniqueName: \"kubernetes.io/projected/3b528ac3-bde2-404d-b02b-9e7a56710af4-kube-api-access-fqnnf\") pod \"3b528ac3-bde2-404d-b02b-9e7a56710af4\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.275852 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-inventory\") pod \"3b528ac3-bde2-404d-b02b-9e7a56710af4\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.275925 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-ssh-key\") pod \"3b528ac3-bde2-404d-b02b-9e7a56710af4\" (UID: \"3b528ac3-bde2-404d-b02b-9e7a56710af4\") " Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.283617 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b528ac3-bde2-404d-b02b-9e7a56710af4-kube-api-access-fqnnf" (OuterVolumeSpecName: "kube-api-access-fqnnf") pod "3b528ac3-bde2-404d-b02b-9e7a56710af4" (UID: "3b528ac3-bde2-404d-b02b-9e7a56710af4"). InnerVolumeSpecName "kube-api-access-fqnnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.304957 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-inventory" (OuterVolumeSpecName: "inventory") pod "3b528ac3-bde2-404d-b02b-9e7a56710af4" (UID: "3b528ac3-bde2-404d-b02b-9e7a56710af4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.313357 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3b528ac3-bde2-404d-b02b-9e7a56710af4" (UID: "3b528ac3-bde2-404d-b02b-9e7a56710af4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.377889 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.377927 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b528ac3-bde2-404d-b02b-9e7a56710af4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.377936 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqnnf\" (UniqueName: \"kubernetes.io/projected/3b528ac3-bde2-404d-b02b-9e7a56710af4-kube-api-access-fqnnf\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.733531 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" event={"ID":"3b528ac3-bde2-404d-b02b-9e7a56710af4","Type":"ContainerDied","Data":"18cc7c33e36eb51f3f41b7ddaf6451c9979588a555f94d72fe3575f00bd4203c"} Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.733598 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18cc7c33e36eb51f3f41b7ddaf6451c9979588a555f94d72fe3575f00bd4203c" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.733682 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.870480 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf"] Nov 24 14:59:24 crc kubenswrapper[4822]: E1124 14:59:24.870894 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b528ac3-bde2-404d-b02b-9e7a56710af4" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.870912 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b528ac3-bde2-404d-b02b-9e7a56710af4" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.871086 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b528ac3-bde2-404d-b02b-9e7a56710af4" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.871816 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.875345 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.875522 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.875626 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.877664 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.884140 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf"] Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.997130 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pjjnf\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.997255 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hcc\" (UniqueName: \"kubernetes.io/projected/9c88d30d-9d50-4018-8ea3-aa480b757eee-kube-api-access-87hcc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pjjnf\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:24 crc kubenswrapper[4822]: I1124 14:59:24.997338 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pjjnf\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:25 crc kubenswrapper[4822]: I1124 14:59:25.098600 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pjjnf\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:25 crc kubenswrapper[4822]: I1124 14:59:25.098749 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hcc\" (UniqueName: \"kubernetes.io/projected/9c88d30d-9d50-4018-8ea3-aa480b757eee-kube-api-access-87hcc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pjjnf\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:25 crc kubenswrapper[4822]: I1124 14:59:25.098895 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pjjnf\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:25 crc kubenswrapper[4822]: I1124 14:59:25.104706 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pjjnf\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:25 crc kubenswrapper[4822]: I1124 14:59:25.106067 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pjjnf\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:25 crc kubenswrapper[4822]: I1124 14:59:25.115417 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hcc\" (UniqueName: \"kubernetes.io/projected/9c88d30d-9d50-4018-8ea3-aa480b757eee-kube-api-access-87hcc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pjjnf\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:25 crc kubenswrapper[4822]: I1124 14:59:25.187283 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 14:59:25 crc kubenswrapper[4822]: W1124 14:59:25.717354 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c88d30d_9d50_4018_8ea3_aa480b757eee.slice/crio-41739b079a6029fc5c264749a502061eae1e236cea6fc96f62805faa8ca8f269 WatchSource:0}: Error finding container 41739b079a6029fc5c264749a502061eae1e236cea6fc96f62805faa8ca8f269: Status 404 returned error can't find the container with id 41739b079a6029fc5c264749a502061eae1e236cea6fc96f62805faa8ca8f269 Nov 24 14:59:25 crc kubenswrapper[4822]: I1124 14:59:25.719461 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf"] Nov 24 14:59:25 crc kubenswrapper[4822]: I1124 14:59:25.744763 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" event={"ID":"9c88d30d-9d50-4018-8ea3-aa480b757eee","Type":"ContainerStarted","Data":"41739b079a6029fc5c264749a502061eae1e236cea6fc96f62805faa8ca8f269"} Nov 24 14:59:26 crc kubenswrapper[4822]: I1124 14:59:26.754878 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" event={"ID":"9c88d30d-9d50-4018-8ea3-aa480b757eee","Type":"ContainerStarted","Data":"559f9cb79bd9e079dc150597f34d24f2aaba4175fdcfe1315617caedcec04612"} Nov 24 14:59:26 crc kubenswrapper[4822]: I1124 14:59:26.779292 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" podStartSLOduration=2.295071729 podStartE2EDuration="2.779272023s" podCreationTimestamp="2025-11-24 14:59:24 +0000 UTC" firstStartedPulling="2025-11-24 14:59:25.720339672 +0000 UTC m=+2402.836980149" lastFinishedPulling="2025-11-24 14:59:26.204539926 +0000 UTC m=+2403.321180443" observedRunningTime="2025-11-24 14:59:26.769754236 +0000 UTC m=+2403.886394733" watchObservedRunningTime="2025-11-24 14:59:26.779272023 +0000 UTC m=+2403.895912510" Nov 24 14:59:37 crc kubenswrapper[4822]: I1124 14:59:37.705657 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:59:37 crc kubenswrapper[4822]: E1124 14:59:37.707009 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 14:59:52 crc kubenswrapper[4822]: I1124 14:59:52.704172 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 14:59:52 crc kubenswrapper[4822]: E1124 14:59:52.704997 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.165703 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f"] Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.167893 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.170510 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.171988 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.186249 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f"] Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.269529 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-config-volume\") pod \"collect-profiles-29399940-twp4f\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.269593 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2llw\" (UniqueName: \"kubernetes.io/projected/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-kube-api-access-h2llw\") pod \"collect-profiles-29399940-twp4f\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.269749 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-secret-volume\") pod \"collect-profiles-29399940-twp4f\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.371621 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-config-volume\") pod \"collect-profiles-29399940-twp4f\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.371959 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2llw\" (UniqueName: \"kubernetes.io/projected/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-kube-api-access-h2llw\") pod \"collect-profiles-29399940-twp4f\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.372085 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-secret-volume\") pod \"collect-profiles-29399940-twp4f\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.372596 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-config-volume\") pod \"collect-profiles-29399940-twp4f\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.377563 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-secret-volume\") pod \"collect-profiles-29399940-twp4f\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.387738 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2llw\" (UniqueName: \"kubernetes.io/projected/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-kube-api-access-h2llw\") pod \"collect-profiles-29399940-twp4f\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.489523 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:00 crc kubenswrapper[4822]: I1124 15:00:00.957428 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f"] Nov 24 15:00:01 crc kubenswrapper[4822]: I1124 15:00:01.321183 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" event={"ID":"96870039-9ba5-4ed9-aa2c-5a4b79c9db51","Type":"ContainerStarted","Data":"b868cb0a3cfa0b9d019d345c53036fb7c6c0722452dcafee3845358b5cf2088b"} Nov 24 15:00:01 crc kubenswrapper[4822]: I1124 15:00:01.321268 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" event={"ID":"96870039-9ba5-4ed9-aa2c-5a4b79c9db51","Type":"ContainerStarted","Data":"ecf8f2ab7705a276af5586865d86c0fe6b69b2dab471be5446778811b4185c51"} Nov 24 15:00:01 crc kubenswrapper[4822]: I1124 15:00:01.350653 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" podStartSLOduration=1.3506313620000001 podStartE2EDuration="1.350631362s" podCreationTimestamp="2025-11-24 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:00:01.336519208 +0000 UTC m=+2438.453159725" watchObservedRunningTime="2025-11-24 15:00:01.350631362 +0000 UTC m=+2438.467271849" Nov 24 15:00:02 crc kubenswrapper[4822]: I1124 15:00:02.332514 4822 generic.go:334] "Generic (PLEG): container finished" podID="96870039-9ba5-4ed9-aa2c-5a4b79c9db51" containerID="b868cb0a3cfa0b9d019d345c53036fb7c6c0722452dcafee3845358b5cf2088b" exitCode=0 Nov 24 15:00:02 crc kubenswrapper[4822]: I1124 15:00:02.332579 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" event={"ID":"96870039-9ba5-4ed9-aa2c-5a4b79c9db51","Type":"ContainerDied","Data":"b868cb0a3cfa0b9d019d345c53036fb7c6c0722452dcafee3845358b5cf2088b"} Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.813178 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.860701 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-config-volume\") pod \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.860915 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-secret-volume\") pod \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.860982 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2llw\" (UniqueName: \"kubernetes.io/projected/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-kube-api-access-h2llw\") pod \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\" (UID: \"96870039-9ba5-4ed9-aa2c-5a4b79c9db51\") " Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.862086 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-config-volume" (OuterVolumeSpecName: "config-volume") pod "96870039-9ba5-4ed9-aa2c-5a4b79c9db51" (UID: "96870039-9ba5-4ed9-aa2c-5a4b79c9db51"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.867638 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-kube-api-access-h2llw" (OuterVolumeSpecName: "kube-api-access-h2llw") pod "96870039-9ba5-4ed9-aa2c-5a4b79c9db51" (UID: "96870039-9ba5-4ed9-aa2c-5a4b79c9db51"). InnerVolumeSpecName "kube-api-access-h2llw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.868321 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "96870039-9ba5-4ed9-aa2c-5a4b79c9db51" (UID: "96870039-9ba5-4ed9-aa2c-5a4b79c9db51"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.963884 4822 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.963926 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2llw\" (UniqueName: \"kubernetes.io/projected/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-kube-api-access-h2llw\") on node \"crc\" DevicePath \"\"" Nov 24 15:00:03 crc kubenswrapper[4822]: I1124 15:00:03.963939 4822 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96870039-9ba5-4ed9-aa2c-5a4b79c9db51-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:00:04 crc kubenswrapper[4822]: I1124 15:00:04.357701 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" event={"ID":"96870039-9ba5-4ed9-aa2c-5a4b79c9db51","Type":"ContainerDied","Data":"ecf8f2ab7705a276af5586865d86c0fe6b69b2dab471be5446778811b4185c51"} Nov 24 15:00:04 crc kubenswrapper[4822]: I1124 15:00:04.358045 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecf8f2ab7705a276af5586865d86c0fe6b69b2dab471be5446778811b4185c51" Nov 24 15:00:04 crc kubenswrapper[4822]: I1124 15:00:04.357788 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-twp4f" Nov 24 15:00:04 crc kubenswrapper[4822]: I1124 15:00:04.435276 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5"] Nov 24 15:00:04 crc kubenswrapper[4822]: I1124 15:00:04.442175 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-j6qx5"] Nov 24 15:00:05 crc kubenswrapper[4822]: I1124 15:00:05.705534 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:00:05 crc kubenswrapper[4822]: E1124 15:00:05.705794 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:00:05 crc kubenswrapper[4822]: I1124 15:00:05.723827 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="979147c0-9b48-4a0f-9506-00b31ca99603" path="/var/lib/kubelet/pods/979147c0-9b48-4a0f-9506-00b31ca99603/volumes" Nov 24 15:00:09 crc kubenswrapper[4822]: I1124 15:00:09.422255 4822 generic.go:334] "Generic (PLEG): container finished" podID="9c88d30d-9d50-4018-8ea3-aa480b757eee" containerID="559f9cb79bd9e079dc150597f34d24f2aaba4175fdcfe1315617caedcec04612" exitCode=0 Nov 24 15:00:09 crc kubenswrapper[4822]: I1124 15:00:09.422371 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" event={"ID":"9c88d30d-9d50-4018-8ea3-aa480b757eee","Type":"ContainerDied","Data":"559f9cb79bd9e079dc150597f34d24f2aaba4175fdcfe1315617caedcec04612"} Nov 24 15:00:10 crc kubenswrapper[4822]: I1124 15:00:10.955490 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.016490 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-inventory\") pod \"9c88d30d-9d50-4018-8ea3-aa480b757eee\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.016747 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87hcc\" (UniqueName: \"kubernetes.io/projected/9c88d30d-9d50-4018-8ea3-aa480b757eee-kube-api-access-87hcc\") pod \"9c88d30d-9d50-4018-8ea3-aa480b757eee\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.019830 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-ssh-key\") pod \"9c88d30d-9d50-4018-8ea3-aa480b757eee\" (UID: \"9c88d30d-9d50-4018-8ea3-aa480b757eee\") " Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.031198 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c88d30d-9d50-4018-8ea3-aa480b757eee-kube-api-access-87hcc" (OuterVolumeSpecName: "kube-api-access-87hcc") pod "9c88d30d-9d50-4018-8ea3-aa480b757eee" (UID: "9c88d30d-9d50-4018-8ea3-aa480b757eee"). InnerVolumeSpecName "kube-api-access-87hcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.056076 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-inventory" (OuterVolumeSpecName: "inventory") pod "9c88d30d-9d50-4018-8ea3-aa480b757eee" (UID: "9c88d30d-9d50-4018-8ea3-aa480b757eee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.068051 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9c88d30d-9d50-4018-8ea3-aa480b757eee" (UID: "9c88d30d-9d50-4018-8ea3-aa480b757eee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.124729 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.124778 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88d30d-9d50-4018-8ea3-aa480b757eee-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.124791 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87hcc\" (UniqueName: \"kubernetes.io/projected/9c88d30d-9d50-4018-8ea3-aa480b757eee-kube-api-access-87hcc\") on node \"crc\" DevicePath \"\"" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.450979 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" event={"ID":"9c88d30d-9d50-4018-8ea3-aa480b757eee","Type":"ContainerDied","Data":"41739b079a6029fc5c264749a502061eae1e236cea6fc96f62805faa8ca8f269"} Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.451019 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41739b079a6029fc5c264749a502061eae1e236cea6fc96f62805faa8ca8f269" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.451019 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pjjnf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.571712 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf"] Nov 24 15:00:11 crc kubenswrapper[4822]: E1124 15:00:11.572180 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c88d30d-9d50-4018-8ea3-aa480b757eee" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.572199 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c88d30d-9d50-4018-8ea3-aa480b757eee" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:00:11 crc kubenswrapper[4822]: E1124 15:00:11.572234 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96870039-9ba5-4ed9-aa2c-5a4b79c9db51" containerName="collect-profiles" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.572240 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="96870039-9ba5-4ed9-aa2c-5a4b79c9db51" containerName="collect-profiles" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.572427 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c88d30d-9d50-4018-8ea3-aa480b757eee" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.572458 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="96870039-9ba5-4ed9-aa2c-5a4b79c9db51" containerName="collect-profiles" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.573267 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.575169 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.575478 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.575866 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.588645 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf"] Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.595623 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.637902 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zprf\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.637996 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zprf\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.638041 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4xfw\" (UniqueName: \"kubernetes.io/projected/8e390d8b-af25-4297-9ca2-e8fe094f7f71-kube-api-access-x4xfw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zprf\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.740009 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zprf\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.740098 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zprf\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.740138 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4xfw\" (UniqueName: \"kubernetes.io/projected/8e390d8b-af25-4297-9ca2-e8fe094f7f71-kube-api-access-x4xfw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zprf\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.744705 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zprf\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.750010 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zprf\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.762477 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4xfw\" (UniqueName: \"kubernetes.io/projected/8e390d8b-af25-4297-9ca2-e8fe094f7f71-kube-api-access-x4xfw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zprf\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:11 crc kubenswrapper[4822]: I1124 15:00:11.916774 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:00:12 crc kubenswrapper[4822]: I1124 15:00:12.536870 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf"] Nov 24 15:00:12 crc kubenswrapper[4822]: W1124 15:00:12.536903 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e390d8b_af25_4297_9ca2_e8fe094f7f71.slice/crio-b2aac61d17398d8bb61361b704b6d3f6233de07d52f9c5ff58b15effead23a40 WatchSource:0}: Error finding container b2aac61d17398d8bb61361b704b6d3f6233de07d52f9c5ff58b15effead23a40: Status 404 returned error can't find the container with id b2aac61d17398d8bb61361b704b6d3f6233de07d52f9c5ff58b15effead23a40 Nov 24 15:00:13 crc kubenswrapper[4822]: I1124 15:00:13.485849 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" event={"ID":"8e390d8b-af25-4297-9ca2-e8fe094f7f71","Type":"ContainerStarted","Data":"b876ab28732dd449808ef18034f83d76d22be56649e8bc24e044180539476471"} Nov 24 15:00:13 crc kubenswrapper[4822]: I1124 15:00:13.486706 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" event={"ID":"8e390d8b-af25-4297-9ca2-e8fe094f7f71","Type":"ContainerStarted","Data":"b2aac61d17398d8bb61361b704b6d3f6233de07d52f9c5ff58b15effead23a40"} Nov 24 15:00:13 crc kubenswrapper[4822]: I1124 15:00:13.510430 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" podStartSLOduration=2.038734209 podStartE2EDuration="2.510401047s" podCreationTimestamp="2025-11-24 15:00:11 +0000 UTC" firstStartedPulling="2025-11-24 15:00:12.539312829 +0000 UTC m=+2449.655953316" lastFinishedPulling="2025-11-24 15:00:13.010979647 +0000 UTC m=+2450.127620154" observedRunningTime="2025-11-24 15:00:13.508770906 +0000 UTC m=+2450.625411463" watchObservedRunningTime="2025-11-24 15:00:13.510401047 +0000 UTC m=+2450.627041534" Nov 24 15:00:19 crc kubenswrapper[4822]: I1124 15:00:19.705954 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:00:19 crc kubenswrapper[4822]: E1124 15:00:19.706901 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:00:29 crc kubenswrapper[4822]: I1124 15:00:29.376977 4822 scope.go:117] "RemoveContainer" containerID="3e0bf0a4ed5ea6616261881f3c0b078da04d750efe64fbda108e4847a7349ea5" Nov 24 15:00:33 crc kubenswrapper[4822]: I1124 15:00:33.718080 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:00:33 crc kubenswrapper[4822]: E1124 15:00:33.719074 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:00:47 crc kubenswrapper[4822]: I1124 15:00:47.705326 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:00:47 crc kubenswrapper[4822]: E1124 15:00:47.706162 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.153625 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29399941-rxk5k"] Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.156006 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.169886 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399941-rxk5k"] Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.291048 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz52x\" (UniqueName: \"kubernetes.io/projected/07506623-c231-4474-9fbf-64be63010543-kube-api-access-rz52x\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.291120 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-fernet-keys\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.291294 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-config-data\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.291375 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-combined-ca-bundle\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.393048 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-combined-ca-bundle\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.393184 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz52x\" (UniqueName: \"kubernetes.io/projected/07506623-c231-4474-9fbf-64be63010543-kube-api-access-rz52x\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.393222 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-fernet-keys\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.393286 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-config-data\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.400964 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-config-data\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.400964 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-fernet-keys\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.402073 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-combined-ca-bundle\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.422923 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz52x\" (UniqueName: \"kubernetes.io/projected/07506623-c231-4474-9fbf-64be63010543-kube-api-access-rz52x\") pod \"keystone-cron-29399941-rxk5k\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.479078 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.704787 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:01:00 crc kubenswrapper[4822]: E1124 15:01:00.705449 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:01:00 crc kubenswrapper[4822]: I1124 15:01:00.950558 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399941-rxk5k"] Nov 24 15:01:01 crc kubenswrapper[4822]: I1124 15:01:01.038341 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399941-rxk5k" event={"ID":"07506623-c231-4474-9fbf-64be63010543","Type":"ContainerStarted","Data":"11712f8d62ef4b2fef81706683bff7112c6bae5477f01bb5aff0f59d02e9bf5e"} Nov 24 15:01:02 crc kubenswrapper[4822]: I1124 15:01:02.055876 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399941-rxk5k" event={"ID":"07506623-c231-4474-9fbf-64be63010543","Type":"ContainerStarted","Data":"3a5d3c36cf524a55337f69fcf28405707effae7ee61c770791018ed5512ec10d"} Nov 24 15:01:02 crc kubenswrapper[4822]: I1124 15:01:02.081694 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29399941-rxk5k" podStartSLOduration=2.0816747 podStartE2EDuration="2.0816747s" podCreationTimestamp="2025-11-24 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:01:02.07627763 +0000 UTC m=+2499.192918137" watchObservedRunningTime="2025-11-24 15:01:02.0816747 +0000 UTC m=+2499.198315187" Nov 24 15:01:04 crc kubenswrapper[4822]: I1124 15:01:04.075076 4822 generic.go:334] "Generic (PLEG): container finished" podID="07506623-c231-4474-9fbf-64be63010543" containerID="3a5d3c36cf524a55337f69fcf28405707effae7ee61c770791018ed5512ec10d" exitCode=0 Nov 24 15:01:04 crc kubenswrapper[4822]: I1124 15:01:04.075113 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399941-rxk5k" event={"ID":"07506623-c231-4474-9fbf-64be63010543","Type":"ContainerDied","Data":"3a5d3c36cf524a55337f69fcf28405707effae7ee61c770791018ed5512ec10d"} Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.644147 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.718660 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-config-data\") pod \"07506623-c231-4474-9fbf-64be63010543\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.718762 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-fernet-keys\") pod \"07506623-c231-4474-9fbf-64be63010543\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.718891 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-combined-ca-bundle\") pod \"07506623-c231-4474-9fbf-64be63010543\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.718916 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz52x\" (UniqueName: \"kubernetes.io/projected/07506623-c231-4474-9fbf-64be63010543-kube-api-access-rz52x\") pod \"07506623-c231-4474-9fbf-64be63010543\" (UID: \"07506623-c231-4474-9fbf-64be63010543\") " Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.724543 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "07506623-c231-4474-9fbf-64be63010543" (UID: "07506623-c231-4474-9fbf-64be63010543"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.734526 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07506623-c231-4474-9fbf-64be63010543-kube-api-access-rz52x" (OuterVolumeSpecName: "kube-api-access-rz52x") pod "07506623-c231-4474-9fbf-64be63010543" (UID: "07506623-c231-4474-9fbf-64be63010543"). InnerVolumeSpecName "kube-api-access-rz52x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.747644 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07506623-c231-4474-9fbf-64be63010543" (UID: "07506623-c231-4474-9fbf-64be63010543"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.777806 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-config-data" (OuterVolumeSpecName: "config-data") pod "07506623-c231-4474-9fbf-64be63010543" (UID: "07506623-c231-4474-9fbf-64be63010543"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.822902 4822 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.822930 4822 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.823113 4822 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07506623-c231-4474-9fbf-64be63010543-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:05 crc kubenswrapper[4822]: I1124 15:01:05.823419 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz52x\" (UniqueName: \"kubernetes.io/projected/07506623-c231-4474-9fbf-64be63010543-kube-api-access-rz52x\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:06 crc kubenswrapper[4822]: I1124 15:01:06.101321 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399941-rxk5k" event={"ID":"07506623-c231-4474-9fbf-64be63010543","Type":"ContainerDied","Data":"11712f8d62ef4b2fef81706683bff7112c6bae5477f01bb5aff0f59d02e9bf5e"} Nov 24 15:01:06 crc kubenswrapper[4822]: I1124 15:01:06.101367 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11712f8d62ef4b2fef81706683bff7112c6bae5477f01bb5aff0f59d02e9bf5e" Nov 24 15:01:06 crc kubenswrapper[4822]: I1124 15:01:06.101436 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399941-rxk5k" Nov 24 15:01:12 crc kubenswrapper[4822]: I1124 15:01:12.171731 4822 generic.go:334] "Generic (PLEG): container finished" podID="8e390d8b-af25-4297-9ca2-e8fe094f7f71" containerID="b876ab28732dd449808ef18034f83d76d22be56649e8bc24e044180539476471" exitCode=0 Nov 24 15:01:12 crc kubenswrapper[4822]: I1124 15:01:12.171818 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" event={"ID":"8e390d8b-af25-4297-9ca2-e8fe094f7f71","Type":"ContainerDied","Data":"b876ab28732dd449808ef18034f83d76d22be56649e8bc24e044180539476471"} Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.675637 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.695662 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-ssh-key\") pod \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.695832 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-inventory\") pod \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.696008 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4xfw\" (UniqueName: \"kubernetes.io/projected/8e390d8b-af25-4297-9ca2-e8fe094f7f71-kube-api-access-x4xfw\") pod \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\" (UID: \"8e390d8b-af25-4297-9ca2-e8fe094f7f71\") " Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.701068 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e390d8b-af25-4297-9ca2-e8fe094f7f71-kube-api-access-x4xfw" (OuterVolumeSpecName: "kube-api-access-x4xfw") pod "8e390d8b-af25-4297-9ca2-e8fe094f7f71" (UID: "8e390d8b-af25-4297-9ca2-e8fe094f7f71"). InnerVolumeSpecName "kube-api-access-x4xfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.719332 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:01:13 crc kubenswrapper[4822]: E1124 15:01:13.719683 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.726744 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-inventory" (OuterVolumeSpecName: "inventory") pod "8e390d8b-af25-4297-9ca2-e8fe094f7f71" (UID: "8e390d8b-af25-4297-9ca2-e8fe094f7f71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.741337 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8e390d8b-af25-4297-9ca2-e8fe094f7f71" (UID: "8e390d8b-af25-4297-9ca2-e8fe094f7f71"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.799286 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.799323 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e390d8b-af25-4297-9ca2-e8fe094f7f71-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:13 crc kubenswrapper[4822]: I1124 15:01:13.799332 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4xfw\" (UniqueName: \"kubernetes.io/projected/8e390d8b-af25-4297-9ca2-e8fe094f7f71-kube-api-access-x4xfw\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.195170 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" event={"ID":"8e390d8b-af25-4297-9ca2-e8fe094f7f71","Type":"ContainerDied","Data":"b2aac61d17398d8bb61361b704b6d3f6233de07d52f9c5ff58b15effead23a40"} Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.195485 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2aac61d17398d8bb61361b704b6d3f6233de07d52f9c5ff58b15effead23a40" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.195325 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zprf" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.286510 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5ldtr"] Nov 24 15:01:14 crc kubenswrapper[4822]: E1124 15:01:14.287111 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e390d8b-af25-4297-9ca2-e8fe094f7f71" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.287132 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e390d8b-af25-4297-9ca2-e8fe094f7f71" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:01:14 crc kubenswrapper[4822]: E1124 15:01:14.287150 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07506623-c231-4474-9fbf-64be63010543" containerName="keystone-cron" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.287159 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="07506623-c231-4474-9fbf-64be63010543" containerName="keystone-cron" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.287415 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="07506623-c231-4474-9fbf-64be63010543" containerName="keystone-cron" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.287444 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e390d8b-af25-4297-9ca2-e8fe094f7f71" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.288428 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.290703 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.290811 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.292849 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.293045 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.297195 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5ldtr"] Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.309553 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7lrp\" (UniqueName: \"kubernetes.io/projected/3e26bb17-0138-493a-90d6-8164ca8367dd-kube-api-access-l7lrp\") pod \"ssh-known-hosts-edpm-deployment-5ldtr\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.309717 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5ldtr\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.309738 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5ldtr\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.410826 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5ldtr\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.410863 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5ldtr\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.410943 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7lrp\" (UniqueName: \"kubernetes.io/projected/3e26bb17-0138-493a-90d6-8164ca8367dd-kube-api-access-l7lrp\") pod \"ssh-known-hosts-edpm-deployment-5ldtr\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.415707 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5ldtr\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.415860 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5ldtr\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.434791 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7lrp\" (UniqueName: \"kubernetes.io/projected/3e26bb17-0138-493a-90d6-8164ca8367dd-kube-api-access-l7lrp\") pod \"ssh-known-hosts-edpm-deployment-5ldtr\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:14 crc kubenswrapper[4822]: I1124 15:01:14.615060 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:15 crc kubenswrapper[4822]: I1124 15:01:15.194606 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5ldtr"] Nov 24 15:01:15 crc kubenswrapper[4822]: I1124 15:01:15.216186 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" event={"ID":"3e26bb17-0138-493a-90d6-8164ca8367dd","Type":"ContainerStarted","Data":"5f826fbfa93b170df860901d4a62885996059dec47cd360822b7e96540f77f2b"} Nov 24 15:01:16 crc kubenswrapper[4822]: I1124 15:01:16.232964 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" event={"ID":"3e26bb17-0138-493a-90d6-8164ca8367dd","Type":"ContainerStarted","Data":"19f382cadf3741359ed3c42cfcc0aa844147ad4faad714bd23462be148a039f8"} Nov 24 15:01:16 crc kubenswrapper[4822]: I1124 15:01:16.255812 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" podStartSLOduration=1.496259619 podStartE2EDuration="2.255795252s" podCreationTimestamp="2025-11-24 15:01:14 +0000 UTC" firstStartedPulling="2025-11-24 15:01:15.204921604 +0000 UTC m=+2512.321562111" lastFinishedPulling="2025-11-24 15:01:15.964457267 +0000 UTC m=+2513.081097744" observedRunningTime="2025-11-24 15:01:16.253430868 +0000 UTC m=+2513.370071365" watchObservedRunningTime="2025-11-24 15:01:16.255795252 +0000 UTC m=+2513.372435739" Nov 24 15:01:24 crc kubenswrapper[4822]: I1124 15:01:24.334055 4822 generic.go:334] "Generic (PLEG): container finished" podID="3e26bb17-0138-493a-90d6-8164ca8367dd" containerID="19f382cadf3741359ed3c42cfcc0aa844147ad4faad714bd23462be148a039f8" exitCode=0 Nov 24 15:01:24 crc kubenswrapper[4822]: I1124 15:01:24.334235 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" event={"ID":"3e26bb17-0138-493a-90d6-8164ca8367dd","Type":"ContainerDied","Data":"19f382cadf3741359ed3c42cfcc0aa844147ad4faad714bd23462be148a039f8"} Nov 24 15:01:25 crc kubenswrapper[4822]: I1124 15:01:25.853545 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.043671 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7lrp\" (UniqueName: \"kubernetes.io/projected/3e26bb17-0138-493a-90d6-8164ca8367dd-kube-api-access-l7lrp\") pod \"3e26bb17-0138-493a-90d6-8164ca8367dd\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.043840 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-inventory-0\") pod \"3e26bb17-0138-493a-90d6-8164ca8367dd\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.044255 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-ssh-key-openstack-edpm-ipam\") pod \"3e26bb17-0138-493a-90d6-8164ca8367dd\" (UID: \"3e26bb17-0138-493a-90d6-8164ca8367dd\") " Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.057266 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e26bb17-0138-493a-90d6-8164ca8367dd-kube-api-access-l7lrp" (OuterVolumeSpecName: "kube-api-access-l7lrp") pod "3e26bb17-0138-493a-90d6-8164ca8367dd" (UID: "3e26bb17-0138-493a-90d6-8164ca8367dd"). InnerVolumeSpecName "kube-api-access-l7lrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.075106 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3e26bb17-0138-493a-90d6-8164ca8367dd" (UID: "3e26bb17-0138-493a-90d6-8164ca8367dd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.080178 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "3e26bb17-0138-493a-90d6-8164ca8367dd" (UID: "3e26bb17-0138-493a-90d6-8164ca8367dd"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.147548 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.147586 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7lrp\" (UniqueName: \"kubernetes.io/projected/3e26bb17-0138-493a-90d6-8164ca8367dd-kube-api-access-l7lrp\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.147597 4822 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3e26bb17-0138-493a-90d6-8164ca8367dd-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.356556 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" event={"ID":"3e26bb17-0138-493a-90d6-8164ca8367dd","Type":"ContainerDied","Data":"5f826fbfa93b170df860901d4a62885996059dec47cd360822b7e96540f77f2b"} Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.356598 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f826fbfa93b170df860901d4a62885996059dec47cd360822b7e96540f77f2b" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.356625 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5ldtr" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.438307 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk"] Nov 24 15:01:26 crc kubenswrapper[4822]: E1124 15:01:26.438820 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e26bb17-0138-493a-90d6-8164ca8367dd" containerName="ssh-known-hosts-edpm-deployment" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.438841 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e26bb17-0138-493a-90d6-8164ca8367dd" containerName="ssh-known-hosts-edpm-deployment" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.439136 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e26bb17-0138-493a-90d6-8164ca8367dd" containerName="ssh-known-hosts-edpm-deployment" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.439914 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.442268 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.442411 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.442610 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.442874 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.452343 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk"] Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.557645 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc427\" (UniqueName: \"kubernetes.io/projected/a385f6bf-f192-408b-9ed4-fe18f7784682-kube-api-access-gc427\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkbkk\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.557937 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkbkk\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.558139 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkbkk\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.659861 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkbkk\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.660313 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc427\" (UniqueName: \"kubernetes.io/projected/a385f6bf-f192-408b-9ed4-fe18f7784682-kube-api-access-gc427\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkbkk\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.660474 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkbkk\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.667852 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkbkk\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.669670 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkbkk\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.688575 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc427\" (UniqueName: \"kubernetes.io/projected/a385f6bf-f192-408b-9ed4-fe18f7784682-kube-api-access-gc427\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkbkk\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:26 crc kubenswrapper[4822]: I1124 15:01:26.766388 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:27 crc kubenswrapper[4822]: I1124 15:01:27.312856 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk"] Nov 24 15:01:27 crc kubenswrapper[4822]: I1124 15:01:27.365772 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" event={"ID":"a385f6bf-f192-408b-9ed4-fe18f7784682","Type":"ContainerStarted","Data":"75af0511ce8938832ebdb152870f8a996a9927c0ac6392f4585ac6d140035f1d"} Nov 24 15:01:28 crc kubenswrapper[4822]: I1124 15:01:28.376498 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" event={"ID":"a385f6bf-f192-408b-9ed4-fe18f7784682","Type":"ContainerStarted","Data":"ffb5c758378ad8f5e25f26269f121b14bbee5f3a820e3c0d163754e421d37efe"} Nov 24 15:01:28 crc kubenswrapper[4822]: I1124 15:01:28.400325 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" podStartSLOduration=2.005945983 podStartE2EDuration="2.400307739s" podCreationTimestamp="2025-11-24 15:01:26 +0000 UTC" firstStartedPulling="2025-11-24 15:01:27.314621066 +0000 UTC m=+2524.431261543" lastFinishedPulling="2025-11-24 15:01:27.708982822 +0000 UTC m=+2524.825623299" observedRunningTime="2025-11-24 15:01:28.399505863 +0000 UTC m=+2525.516146360" watchObservedRunningTime="2025-11-24 15:01:28.400307739 +0000 UTC m=+2525.516948226" Nov 24 15:01:28 crc kubenswrapper[4822]: I1124 15:01:28.707142 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:01:28 crc kubenswrapper[4822]: E1124 15:01:28.707750 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:01:37 crc kubenswrapper[4822]: I1124 15:01:37.483292 4822 generic.go:334] "Generic (PLEG): container finished" podID="a385f6bf-f192-408b-9ed4-fe18f7784682" containerID="ffb5c758378ad8f5e25f26269f121b14bbee5f3a820e3c0d163754e421d37efe" exitCode=0 Nov 24 15:01:37 crc kubenswrapper[4822]: I1124 15:01:37.483400 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" event={"ID":"a385f6bf-f192-408b-9ed4-fe18f7784682","Type":"ContainerDied","Data":"ffb5c758378ad8f5e25f26269f121b14bbee5f3a820e3c0d163754e421d37efe"} Nov 24 15:01:38 crc kubenswrapper[4822]: I1124 15:01:38.985625 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.059822 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-inventory\") pod \"a385f6bf-f192-408b-9ed4-fe18f7784682\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.060422 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc427\" (UniqueName: \"kubernetes.io/projected/a385f6bf-f192-408b-9ed4-fe18f7784682-kube-api-access-gc427\") pod \"a385f6bf-f192-408b-9ed4-fe18f7784682\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.060478 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-ssh-key\") pod \"a385f6bf-f192-408b-9ed4-fe18f7784682\" (UID: \"a385f6bf-f192-408b-9ed4-fe18f7784682\") " Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.076548 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a385f6bf-f192-408b-9ed4-fe18f7784682-kube-api-access-gc427" (OuterVolumeSpecName: "kube-api-access-gc427") pod "a385f6bf-f192-408b-9ed4-fe18f7784682" (UID: "a385f6bf-f192-408b-9ed4-fe18f7784682"). InnerVolumeSpecName "kube-api-access-gc427". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.099463 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a385f6bf-f192-408b-9ed4-fe18f7784682" (UID: "a385f6bf-f192-408b-9ed4-fe18f7784682"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.111718 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-inventory" (OuterVolumeSpecName: "inventory") pod "a385f6bf-f192-408b-9ed4-fe18f7784682" (UID: "a385f6bf-f192-408b-9ed4-fe18f7784682"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.164328 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc427\" (UniqueName: \"kubernetes.io/projected/a385f6bf-f192-408b-9ed4-fe18f7784682-kube-api-access-gc427\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.164370 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.164382 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a385f6bf-f192-408b-9ed4-fe18f7784682-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.521031 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" event={"ID":"a385f6bf-f192-408b-9ed4-fe18f7784682","Type":"ContainerDied","Data":"75af0511ce8938832ebdb152870f8a996a9927c0ac6392f4585ac6d140035f1d"} Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.521271 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75af0511ce8938832ebdb152870f8a996a9927c0ac6392f4585ac6d140035f1d" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.521176 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkbkk" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.627655 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s"] Nov 24 15:01:39 crc kubenswrapper[4822]: E1124 15:01:39.628225 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a385f6bf-f192-408b-9ed4-fe18f7784682" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.628249 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="a385f6bf-f192-408b-9ed4-fe18f7784682" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.628496 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="a385f6bf-f192-408b-9ed4-fe18f7784682" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.629414 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.631255 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.631477 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.631496 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.631916 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.659424 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s"] Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.674675 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8p9l\" (UniqueName: \"kubernetes.io/projected/ea1150ac-561e-4776-a2b4-233a5d4189be-kube-api-access-m8p9l\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.674751 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.674924 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.775871 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8p9l\" (UniqueName: \"kubernetes.io/projected/ea1150ac-561e-4776-a2b4-233a5d4189be-kube-api-access-m8p9l\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.775959 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.776927 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.782148 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.782243 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.792515 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8p9l\" (UniqueName: \"kubernetes.io/projected/ea1150ac-561e-4776-a2b4-233a5d4189be-kube-api-access-m8p9l\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:39 crc kubenswrapper[4822]: I1124 15:01:39.957937 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:40 crc kubenswrapper[4822]: I1124 15:01:40.568779 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s"] Nov 24 15:01:40 crc kubenswrapper[4822]: W1124 15:01:40.576681 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea1150ac_561e_4776_a2b4_233a5d4189be.slice/crio-6995e5599bb9f7df3b69ac3f9266098dd6a697b1cfecabdff3d6167830b7a4c9 WatchSource:0}: Error finding container 6995e5599bb9f7df3b69ac3f9266098dd6a697b1cfecabdff3d6167830b7a4c9: Status 404 returned error can't find the container with id 6995e5599bb9f7df3b69ac3f9266098dd6a697b1cfecabdff3d6167830b7a4c9 Nov 24 15:01:40 crc kubenswrapper[4822]: I1124 15:01:40.704723 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:01:40 crc kubenswrapper[4822]: E1124 15:01:40.705014 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:01:41 crc kubenswrapper[4822]: I1124 15:01:41.544525 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" event={"ID":"ea1150ac-561e-4776-a2b4-233a5d4189be","Type":"ContainerStarted","Data":"f163f0eb5bb1332c4aa45341a0c29fbceb3b08cf2159ba4fabf1a894b9636d19"} Nov 24 15:01:41 crc kubenswrapper[4822]: I1124 15:01:41.544921 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" event={"ID":"ea1150ac-561e-4776-a2b4-233a5d4189be","Type":"ContainerStarted","Data":"6995e5599bb9f7df3b69ac3f9266098dd6a697b1cfecabdff3d6167830b7a4c9"} Nov 24 15:01:41 crc kubenswrapper[4822]: I1124 15:01:41.570025 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" podStartSLOduration=2.177881821 podStartE2EDuration="2.569997535s" podCreationTimestamp="2025-11-24 15:01:39 +0000 UTC" firstStartedPulling="2025-11-24 15:01:40.581182623 +0000 UTC m=+2537.697823110" lastFinishedPulling="2025-11-24 15:01:40.973298307 +0000 UTC m=+2538.089938824" observedRunningTime="2025-11-24 15:01:41.558596997 +0000 UTC m=+2538.675237474" watchObservedRunningTime="2025-11-24 15:01:41.569997535 +0000 UTC m=+2538.686638022" Nov 24 15:01:45 crc kubenswrapper[4822]: I1124 15:01:45.067361 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-kks22"] Nov 24 15:01:45 crc kubenswrapper[4822]: I1124 15:01:45.084573 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-kks22"] Nov 24 15:01:45 crc kubenswrapper[4822]: I1124 15:01:45.721135 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f64cdb18-1ccf-46b4-bea2-9df357833a51" path="/var/lib/kubelet/pods/f64cdb18-1ccf-46b4-bea2-9df357833a51/volumes" Nov 24 15:01:52 crc kubenswrapper[4822]: I1124 15:01:52.049033 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-v5pw2"] Nov 24 15:01:52 crc kubenswrapper[4822]: I1124 15:01:52.061294 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-v5pw2"] Nov 24 15:01:52 crc kubenswrapper[4822]: I1124 15:01:52.677570 4822 generic.go:334] "Generic (PLEG): container finished" podID="ea1150ac-561e-4776-a2b4-233a5d4189be" containerID="f163f0eb5bb1332c4aa45341a0c29fbceb3b08cf2159ba4fabf1a894b9636d19" exitCode=0 Nov 24 15:01:52 crc kubenswrapper[4822]: I1124 15:01:52.677644 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" event={"ID":"ea1150ac-561e-4776-a2b4-233a5d4189be","Type":"ContainerDied","Data":"f163f0eb5bb1332c4aa45341a0c29fbceb3b08cf2159ba4fabf1a894b9636d19"} Nov 24 15:01:53 crc kubenswrapper[4822]: I1124 15:01:53.727739 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ed26986-cc75-4e38-b2ab-53b839cbb8d6" path="/var/lib/kubelet/pods/9ed26986-cc75-4e38-b2ab-53b839cbb8d6/volumes" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.287401 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.465882 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8p9l\" (UniqueName: \"kubernetes.io/projected/ea1150ac-561e-4776-a2b4-233a5d4189be-kube-api-access-m8p9l\") pod \"ea1150ac-561e-4776-a2b4-233a5d4189be\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.466356 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-ssh-key\") pod \"ea1150ac-561e-4776-a2b4-233a5d4189be\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.466681 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-inventory\") pod \"ea1150ac-561e-4776-a2b4-233a5d4189be\" (UID: \"ea1150ac-561e-4776-a2b4-233a5d4189be\") " Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.472904 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea1150ac-561e-4776-a2b4-233a5d4189be-kube-api-access-m8p9l" (OuterVolumeSpecName: "kube-api-access-m8p9l") pod "ea1150ac-561e-4776-a2b4-233a5d4189be" (UID: "ea1150ac-561e-4776-a2b4-233a5d4189be"). InnerVolumeSpecName "kube-api-access-m8p9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.502521 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-inventory" (OuterVolumeSpecName: "inventory") pod "ea1150ac-561e-4776-a2b4-233a5d4189be" (UID: "ea1150ac-561e-4776-a2b4-233a5d4189be"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.513164 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ea1150ac-561e-4776-a2b4-233a5d4189be" (UID: "ea1150ac-561e-4776-a2b4-233a5d4189be"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.569108 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.569132 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea1150ac-561e-4776-a2b4-233a5d4189be-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.569142 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8p9l\" (UniqueName: \"kubernetes.io/projected/ea1150ac-561e-4776-a2b4-233a5d4189be-kube-api-access-m8p9l\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.701283 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" event={"ID":"ea1150ac-561e-4776-a2b4-233a5d4189be","Type":"ContainerDied","Data":"6995e5599bb9f7df3b69ac3f9266098dd6a697b1cfecabdff3d6167830b7a4c9"} Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.701343 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6995e5599bb9f7df3b69ac3f9266098dd6a697b1cfecabdff3d6167830b7a4c9" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.701369 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.705546 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.799225 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl"] Nov 24 15:01:54 crc kubenswrapper[4822]: E1124 15:01:54.800824 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1150ac-561e-4776-a2b4-233a5d4189be" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.800928 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1150ac-561e-4776-a2b4-233a5d4189be" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.803518 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea1150ac-561e-4776-a2b4-233a5d4189be" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.804721 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.808320 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.808522 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.808921 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.809039 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.809085 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.809263 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.809411 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.809585 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.822473 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl"] Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.975910 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976330 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976374 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976431 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976483 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976529 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976586 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976632 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976690 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976723 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976775 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976848 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.976935 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dszwv\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-kube-api-access-dszwv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:54 crc kubenswrapper[4822]: I1124 15:01:54.977006 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.078647 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.078732 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.078804 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dszwv\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-kube-api-access-dszwv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.078851 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.078937 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.079009 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.079042 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.079089 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.079135 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.079178 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.079255 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.079296 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.079346 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.079382 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.085760 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.086077 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.086912 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.087670 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.087686 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.087760 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.088970 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.089307 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.089670 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.090004 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.090365 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.091798 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.092791 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.105553 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dszwv\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-kube-api-access-dszwv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sttpl\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.173764 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.743816 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"883d279d8005f9e92a76f4480790753dc5091bbb3ea64a7c08767b819a41a4f8"} Nov 24 15:01:55 crc kubenswrapper[4822]: I1124 15:01:55.811375 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl"] Nov 24 15:01:56 crc kubenswrapper[4822]: I1124 15:01:56.747197 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" event={"ID":"1eca33a0-77b5-42a6-9846-ecb6a0582a87","Type":"ContainerStarted","Data":"00aeb686e04472b598db51a81bd9499f704cb22e67eacbd75eb21f2bd450fde8"} Nov 24 15:01:56 crc kubenswrapper[4822]: I1124 15:01:56.747911 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" event={"ID":"1eca33a0-77b5-42a6-9846-ecb6a0582a87","Type":"ContainerStarted","Data":"9606add6ca20aeecd93feeb42e87771006bc6a5dc10ba1a25b9e662b5f297686"} Nov 24 15:01:56 crc kubenswrapper[4822]: I1124 15:01:56.789963 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" podStartSLOduration=2.351682358 podStartE2EDuration="2.789938437s" podCreationTimestamp="2025-11-24 15:01:54 +0000 UTC" firstStartedPulling="2025-11-24 15:01:55.824653774 +0000 UTC m=+2552.941294281" lastFinishedPulling="2025-11-24 15:01:56.262909863 +0000 UTC m=+2553.379550360" observedRunningTime="2025-11-24 15:01:56.783259968 +0000 UTC m=+2553.899900485" watchObservedRunningTime="2025-11-24 15:01:56.789938437 +0000 UTC m=+2553.906578944" Nov 24 15:02:29 crc kubenswrapper[4822]: I1124 15:02:29.502529 4822 scope.go:117] "RemoveContainer" containerID="6fe9ae59092417bc7ce7646da5ccc38209363d4e24f04c4de0f4349ceac307fc" Nov 24 15:02:29 crc kubenswrapper[4822]: I1124 15:02:29.557412 4822 scope.go:117] "RemoveContainer" containerID="03c9e064b8818806a2256646070656f3060141f999369cc7d12aa82bd1574aac" Nov 24 15:02:41 crc kubenswrapper[4822]: I1124 15:02:41.240456 4822 generic.go:334] "Generic (PLEG): container finished" podID="1eca33a0-77b5-42a6-9846-ecb6a0582a87" containerID="00aeb686e04472b598db51a81bd9499f704cb22e67eacbd75eb21f2bd450fde8" exitCode=0 Nov 24 15:02:41 crc kubenswrapper[4822]: I1124 15:02:41.240566 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" event={"ID":"1eca33a0-77b5-42a6-9846-ecb6a0582a87","Type":"ContainerDied","Data":"00aeb686e04472b598db51a81bd9499f704cb22e67eacbd75eb21f2bd450fde8"} Nov 24 15:02:42 crc kubenswrapper[4822]: I1124 15:02:42.866327 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:02:42 crc kubenswrapper[4822]: I1124 15:02:42.999459 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:42 crc kubenswrapper[4822]: I1124 15:02:42.999519 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:42 crc kubenswrapper[4822]: I1124 15:02:42.999545 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:42 crc kubenswrapper[4822]: I1124 15:02:42.999588 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-nova-combined-ca-bundle\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:42 crc kubenswrapper[4822]: I1124 15:02:42.999616 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-ovn-default-certs-0\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:42.999699 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dszwv\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-kube-api-access-dszwv\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:42.999732 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-repo-setup-combined-ca-bundle\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:42.999775 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ssh-key\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.000004 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-inventory\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.000063 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-libvirt-combined-ca-bundle\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.000099 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-telemetry-combined-ca-bundle\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.000146 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-bootstrap-combined-ca-bundle\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.000175 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ovn-combined-ca-bundle\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.000228 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-neutron-metadata-combined-ca-bundle\") pod \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\" (UID: \"1eca33a0-77b5-42a6-9846-ecb6a0582a87\") " Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.007336 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.009373 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.012374 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-kube-api-access-dszwv" (OuterVolumeSpecName: "kube-api-access-dszwv") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "kube-api-access-dszwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.012488 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.012842 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.013110 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.014293 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.014336 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.014359 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.014548 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.017253 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.027044 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.038128 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-inventory" (OuterVolumeSpecName: "inventory") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.046674 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1eca33a0-77b5-42a6-9846-ecb6a0582a87" (UID: "1eca33a0-77b5-42a6-9846-ecb6a0582a87"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108599 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108654 4822 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108674 4822 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108689 4822 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108704 4822 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108718 4822 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108733 4822 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108750 4822 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108787 4822 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108801 4822 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108814 4822 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108829 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dszwv\" (UniqueName: \"kubernetes.io/projected/1eca33a0-77b5-42a6-9846-ecb6a0582a87-kube-api-access-dszwv\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108845 4822 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.108857 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eca33a0-77b5-42a6-9846-ecb6a0582a87-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.303411 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" event={"ID":"1eca33a0-77b5-42a6-9846-ecb6a0582a87","Type":"ContainerDied","Data":"9606add6ca20aeecd93feeb42e87771006bc6a5dc10ba1a25b9e662b5f297686"} Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.303754 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9606add6ca20aeecd93feeb42e87771006bc6a5dc10ba1a25b9e662b5f297686" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.303995 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sttpl" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.376087 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt"] Nov 24 15:02:43 crc kubenswrapper[4822]: E1124 15:02:43.376514 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eca33a0-77b5-42a6-9846-ecb6a0582a87" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.376537 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eca33a0-77b5-42a6-9846-ecb6a0582a87" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.376758 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eca33a0-77b5-42a6-9846-ecb6a0582a87" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.377566 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.387101 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.387194 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.387259 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.387443 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.387465 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.414770 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt"] Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.423071 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.423370 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.423513 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvmmw\" (UniqueName: \"kubernetes.io/projected/b90073ee-ca05-4991-a028-96fa152f8a71-kube-api-access-lvmmw\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.423648 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.423745 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b90073ee-ca05-4991-a028-96fa152f8a71-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.525701 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.525772 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvmmw\" (UniqueName: \"kubernetes.io/projected/b90073ee-ca05-4991-a028-96fa152f8a71-kube-api-access-lvmmw\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.525821 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b90073ee-ca05-4991-a028-96fa152f8a71-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.525838 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.525940 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.527395 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b90073ee-ca05-4991-a028-96fa152f8a71-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.530449 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.534684 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.535785 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.544955 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvmmw\" (UniqueName: \"kubernetes.io/projected/b90073ee-ca05-4991-a028-96fa152f8a71-kube-api-access-lvmmw\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jxppt\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:43 crc kubenswrapper[4822]: I1124 15:02:43.712893 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:02:44 crc kubenswrapper[4822]: I1124 15:02:44.518079 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt"] Nov 24 15:02:44 crc kubenswrapper[4822]: W1124 15:02:44.523823 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb90073ee_ca05_4991_a028_96fa152f8a71.slice/crio-46fb26feb39c077b3626906279bdac672fce2bb5c8716115539984a522304532 WatchSource:0}: Error finding container 46fb26feb39c077b3626906279bdac672fce2bb5c8716115539984a522304532: Status 404 returned error can't find the container with id 46fb26feb39c077b3626906279bdac672fce2bb5c8716115539984a522304532 Nov 24 15:02:45 crc kubenswrapper[4822]: I1124 15:02:45.326574 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" event={"ID":"b90073ee-ca05-4991-a028-96fa152f8a71","Type":"ContainerStarted","Data":"88d5957985a449ed9598c747ae7fe1474233fd292c56fa0b582c147cc6207c99"} Nov 24 15:02:45 crc kubenswrapper[4822]: I1124 15:02:45.326637 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" event={"ID":"b90073ee-ca05-4991-a028-96fa152f8a71","Type":"ContainerStarted","Data":"46fb26feb39c077b3626906279bdac672fce2bb5c8716115539984a522304532"} Nov 24 15:02:45 crc kubenswrapper[4822]: I1124 15:02:45.346061 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" podStartSLOduration=1.8276276949999999 podStartE2EDuration="2.346043729s" podCreationTimestamp="2025-11-24 15:02:43 +0000 UTC" firstStartedPulling="2025-11-24 15:02:44.526718992 +0000 UTC m=+2601.643359479" lastFinishedPulling="2025-11-24 15:02:45.045135036 +0000 UTC m=+2602.161775513" observedRunningTime="2025-11-24 15:02:45.34413735 +0000 UTC m=+2602.460777897" watchObservedRunningTime="2025-11-24 15:02:45.346043729 +0000 UTC m=+2602.462684206" Nov 24 15:04:00 crc kubenswrapper[4822]: I1124 15:04:00.205616 4822 generic.go:334] "Generic (PLEG): container finished" podID="b90073ee-ca05-4991-a028-96fa152f8a71" containerID="88d5957985a449ed9598c747ae7fe1474233fd292c56fa0b582c147cc6207c99" exitCode=0 Nov 24 15:04:00 crc kubenswrapper[4822]: I1124 15:04:00.205766 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" event={"ID":"b90073ee-ca05-4991-a028-96fa152f8a71","Type":"ContainerDied","Data":"88d5957985a449ed9598c747ae7fe1474233fd292c56fa0b582c147cc6207c99"} Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.716852 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.893137 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvmmw\" (UniqueName: \"kubernetes.io/projected/b90073ee-ca05-4991-a028-96fa152f8a71-kube-api-access-lvmmw\") pod \"b90073ee-ca05-4991-a028-96fa152f8a71\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.893282 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ssh-key\") pod \"b90073ee-ca05-4991-a028-96fa152f8a71\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.893399 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ovn-combined-ca-bundle\") pod \"b90073ee-ca05-4991-a028-96fa152f8a71\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.893456 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-inventory\") pod \"b90073ee-ca05-4991-a028-96fa152f8a71\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.893490 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b90073ee-ca05-4991-a028-96fa152f8a71-ovncontroller-config-0\") pod \"b90073ee-ca05-4991-a028-96fa152f8a71\" (UID: \"b90073ee-ca05-4991-a028-96fa152f8a71\") " Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.911497 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b90073ee-ca05-4991-a028-96fa152f8a71" (UID: "b90073ee-ca05-4991-a028-96fa152f8a71"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.928906 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b90073ee-ca05-4991-a028-96fa152f8a71-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "b90073ee-ca05-4991-a028-96fa152f8a71" (UID: "b90073ee-ca05-4991-a028-96fa152f8a71"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.929399 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b90073ee-ca05-4991-a028-96fa152f8a71-kube-api-access-lvmmw" (OuterVolumeSpecName: "kube-api-access-lvmmw") pod "b90073ee-ca05-4991-a028-96fa152f8a71" (UID: "b90073ee-ca05-4991-a028-96fa152f8a71"). InnerVolumeSpecName "kube-api-access-lvmmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.963291 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-inventory" (OuterVolumeSpecName: "inventory") pod "b90073ee-ca05-4991-a028-96fa152f8a71" (UID: "b90073ee-ca05-4991-a028-96fa152f8a71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.970350 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b90073ee-ca05-4991-a028-96fa152f8a71" (UID: "b90073ee-ca05-4991-a028-96fa152f8a71"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.996471 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.996508 4822 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b90073ee-ca05-4991-a028-96fa152f8a71-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.996521 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvmmw\" (UniqueName: \"kubernetes.io/projected/b90073ee-ca05-4991-a028-96fa152f8a71-kube-api-access-lvmmw\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.996529 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:01 crc kubenswrapper[4822]: I1124 15:04:01.996540 4822 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90073ee-ca05-4991-a028-96fa152f8a71-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.227366 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" event={"ID":"b90073ee-ca05-4991-a028-96fa152f8a71","Type":"ContainerDied","Data":"46fb26feb39c077b3626906279bdac672fce2bb5c8716115539984a522304532"} Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.227584 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46fb26feb39c077b3626906279bdac672fce2bb5c8716115539984a522304532" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.227439 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jxppt" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.320241 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc"] Nov 24 15:04:02 crc kubenswrapper[4822]: E1124 15:04:02.320627 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90073ee-ca05-4991-a028-96fa152f8a71" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.320643 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90073ee-ca05-4991-a028-96fa152f8a71" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.320828 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90073ee-ca05-4991-a028-96fa152f8a71" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.321520 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.332733 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc"] Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.334050 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.334362 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.335552 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.335813 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.335994 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.338135 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.403233 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.403329 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.403380 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.403441 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.403609 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qb7r\" (UniqueName: \"kubernetes.io/projected/f6f36194-152e-41f6-9fa9-fd0f7b279843-kube-api-access-5qb7r\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.403868 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.505555 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.505628 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.505678 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.505716 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.505763 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.505800 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qb7r\" (UniqueName: \"kubernetes.io/projected/f6f36194-152e-41f6-9fa9-fd0f7b279843-kube-api-access-5qb7r\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.510478 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.510489 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.510489 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.511641 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.513377 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.530189 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qb7r\" (UniqueName: \"kubernetes.io/projected/f6f36194-152e-41f6-9fa9-fd0f7b279843-kube-api-access-5qb7r\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:02 crc kubenswrapper[4822]: I1124 15:04:02.644156 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:04:03 crc kubenswrapper[4822]: I1124 15:04:03.213949 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:04:03 crc kubenswrapper[4822]: I1124 15:04:03.214405 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc"] Nov 24 15:04:03 crc kubenswrapper[4822]: I1124 15:04:03.237102 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" event={"ID":"f6f36194-152e-41f6-9fa9-fd0f7b279843","Type":"ContainerStarted","Data":"f86b03d4df201ea4ffadf77544259f551188f64d08c4dbbbd0746bfd77c58fa4"} Nov 24 15:04:04 crc kubenswrapper[4822]: I1124 15:04:04.255221 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" event={"ID":"f6f36194-152e-41f6-9fa9-fd0f7b279843","Type":"ContainerStarted","Data":"a5f4a2526f20d25f29c8f13ae3dc289a8633023fb543e385d3f958977e053aaa"} Nov 24 15:04:04 crc kubenswrapper[4822]: I1124 15:04:04.269095 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" podStartSLOduration=1.714960677 podStartE2EDuration="2.26907936s" podCreationTimestamp="2025-11-24 15:04:02 +0000 UTC" firstStartedPulling="2025-11-24 15:04:03.213677881 +0000 UTC m=+2680.330318358" lastFinishedPulling="2025-11-24 15:04:03.767796554 +0000 UTC m=+2680.884437041" observedRunningTime="2025-11-24 15:04:04.265324332 +0000 UTC m=+2681.381964809" watchObservedRunningTime="2025-11-24 15:04:04.26907936 +0000 UTC m=+2681.385719837" Nov 24 15:04:11 crc kubenswrapper[4822]: I1124 15:04:11.577538 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:04:11 crc kubenswrapper[4822]: I1124 15:04:11.577998 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.205562 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ppvxt"] Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.215483 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.230322 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ppvxt"] Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.287733 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-catalog-content\") pod \"certified-operators-ppvxt\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.287838 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrgfd\" (UniqueName: \"kubernetes.io/projected/bda41bb8-3959-4af5-8d6f-0168d14e76f0-kube-api-access-vrgfd\") pod \"certified-operators-ppvxt\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.287886 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-utilities\") pod \"certified-operators-ppvxt\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.389544 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-utilities\") pod \"certified-operators-ppvxt\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.389744 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-catalog-content\") pod \"certified-operators-ppvxt\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.389782 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrgfd\" (UniqueName: \"kubernetes.io/projected/bda41bb8-3959-4af5-8d6f-0168d14e76f0-kube-api-access-vrgfd\") pod \"certified-operators-ppvxt\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.389940 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-utilities\") pod \"certified-operators-ppvxt\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.390137 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-catalog-content\") pod \"certified-operators-ppvxt\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.409662 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrgfd\" (UniqueName: \"kubernetes.io/projected/bda41bb8-3959-4af5-8d6f-0168d14e76f0-kube-api-access-vrgfd\") pod \"certified-operators-ppvxt\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:36 crc kubenswrapper[4822]: I1124 15:04:36.582534 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:37 crc kubenswrapper[4822]: I1124 15:04:37.242905 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ppvxt"] Nov 24 15:04:37 crc kubenswrapper[4822]: I1124 15:04:37.619850 4822 generic.go:334] "Generic (PLEG): container finished" podID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerID="fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9" exitCode=0 Nov 24 15:04:37 crc kubenswrapper[4822]: I1124 15:04:37.619925 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ppvxt" event={"ID":"bda41bb8-3959-4af5-8d6f-0168d14e76f0","Type":"ContainerDied","Data":"fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9"} Nov 24 15:04:37 crc kubenswrapper[4822]: I1124 15:04:37.620164 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ppvxt" event={"ID":"bda41bb8-3959-4af5-8d6f-0168d14e76f0","Type":"ContainerStarted","Data":"59ccf13f1ef712edd1c4ad27ea05d5c71ad7e4e70122c622f79a000f8d339f10"} Nov 24 15:04:38 crc kubenswrapper[4822]: I1124 15:04:38.647079 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ppvxt" event={"ID":"bda41bb8-3959-4af5-8d6f-0168d14e76f0","Type":"ContainerStarted","Data":"e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba"} Nov 24 15:04:40 crc kubenswrapper[4822]: I1124 15:04:40.668936 4822 generic.go:334] "Generic (PLEG): container finished" podID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerID="e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba" exitCode=0 Nov 24 15:04:40 crc kubenswrapper[4822]: I1124 15:04:40.669020 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ppvxt" event={"ID":"bda41bb8-3959-4af5-8d6f-0168d14e76f0","Type":"ContainerDied","Data":"e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba"} Nov 24 15:04:41 crc kubenswrapper[4822]: I1124 15:04:41.578411 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:04:41 crc kubenswrapper[4822]: I1124 15:04:41.578930 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:04:41 crc kubenswrapper[4822]: I1124 15:04:41.682285 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ppvxt" event={"ID":"bda41bb8-3959-4af5-8d6f-0168d14e76f0","Type":"ContainerStarted","Data":"df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff"} Nov 24 15:04:41 crc kubenswrapper[4822]: I1124 15:04:41.703125 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ppvxt" podStartSLOduration=2.227735245 podStartE2EDuration="5.703107915s" podCreationTimestamp="2025-11-24 15:04:36 +0000 UTC" firstStartedPulling="2025-11-24 15:04:37.621768737 +0000 UTC m=+2714.738409214" lastFinishedPulling="2025-11-24 15:04:41.097141397 +0000 UTC m=+2718.213781884" observedRunningTime="2025-11-24 15:04:41.697888832 +0000 UTC m=+2718.814529329" watchObservedRunningTime="2025-11-24 15:04:41.703107915 +0000 UTC m=+2718.819748382" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.140037 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xgxk8"] Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.143095 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.154704 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgxk8"] Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.204029 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-utilities\") pod \"community-operators-xgxk8\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.204192 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72k62\" (UniqueName: \"kubernetes.io/projected/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-kube-api-access-72k62\") pod \"community-operators-xgxk8\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.204264 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-catalog-content\") pod \"community-operators-xgxk8\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.307797 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72k62\" (UniqueName: \"kubernetes.io/projected/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-kube-api-access-72k62\") pod \"community-operators-xgxk8\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.307864 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-catalog-content\") pod \"community-operators-xgxk8\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.308032 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-utilities\") pod \"community-operators-xgxk8\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.308477 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-catalog-content\") pod \"community-operators-xgxk8\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.308494 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-utilities\") pod \"community-operators-xgxk8\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.327016 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72k62\" (UniqueName: \"kubernetes.io/projected/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-kube-api-access-72k62\") pod \"community-operators-xgxk8\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.462507 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.583794 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.583848 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.706390 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:46 crc kubenswrapper[4822]: I1124 15:04:46.795888 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:47 crc kubenswrapper[4822]: I1124 15:04:47.055569 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgxk8"] Nov 24 15:04:47 crc kubenswrapper[4822]: I1124 15:04:47.749012 4822 generic.go:334] "Generic (PLEG): container finished" podID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerID="8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8" exitCode=0 Nov 24 15:04:47 crc kubenswrapper[4822]: I1124 15:04:47.749058 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgxk8" event={"ID":"c812e1c0-612a-43ed-a767-0da7cf5fa9e3","Type":"ContainerDied","Data":"8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8"} Nov 24 15:04:47 crc kubenswrapper[4822]: I1124 15:04:47.749517 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgxk8" event={"ID":"c812e1c0-612a-43ed-a767-0da7cf5fa9e3","Type":"ContainerStarted","Data":"e6033072bf6c0efea3a78fbcbec8e5d7bb35c0325baff9cc04d9494bc9366eb5"} Nov 24 15:04:48 crc kubenswrapper[4822]: I1124 15:04:48.764325 4822 generic.go:334] "Generic (PLEG): container finished" podID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerID="28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913" exitCode=0 Nov 24 15:04:48 crc kubenswrapper[4822]: I1124 15:04:48.764388 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgxk8" event={"ID":"c812e1c0-612a-43ed-a767-0da7cf5fa9e3","Type":"ContainerDied","Data":"28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913"} Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.116792 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ppvxt"] Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.117462 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ppvxt" podUID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerName="registry-server" containerID="cri-o://df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff" gracePeriod=2 Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.634378 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.743563 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-catalog-content\") pod \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.743809 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrgfd\" (UniqueName: \"kubernetes.io/projected/bda41bb8-3959-4af5-8d6f-0168d14e76f0-kube-api-access-vrgfd\") pod \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.743868 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-utilities\") pod \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\" (UID: \"bda41bb8-3959-4af5-8d6f-0168d14e76f0\") " Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.745010 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-utilities" (OuterVolumeSpecName: "utilities") pod "bda41bb8-3959-4af5-8d6f-0168d14e76f0" (UID: "bda41bb8-3959-4af5-8d6f-0168d14e76f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.754448 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bda41bb8-3959-4af5-8d6f-0168d14e76f0-kube-api-access-vrgfd" (OuterVolumeSpecName: "kube-api-access-vrgfd") pod "bda41bb8-3959-4af5-8d6f-0168d14e76f0" (UID: "bda41bb8-3959-4af5-8d6f-0168d14e76f0"). InnerVolumeSpecName "kube-api-access-vrgfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.776604 4822 generic.go:334] "Generic (PLEG): container finished" podID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerID="df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff" exitCode=0 Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.776681 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ppvxt" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.776732 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ppvxt" event={"ID":"bda41bb8-3959-4af5-8d6f-0168d14e76f0","Type":"ContainerDied","Data":"df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff"} Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.776771 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ppvxt" event={"ID":"bda41bb8-3959-4af5-8d6f-0168d14e76f0","Type":"ContainerDied","Data":"59ccf13f1ef712edd1c4ad27ea05d5c71ad7e4e70122c622f79a000f8d339f10"} Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.776796 4822 scope.go:117] "RemoveContainer" containerID="df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.780651 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgxk8" event={"ID":"c812e1c0-612a-43ed-a767-0da7cf5fa9e3","Type":"ContainerStarted","Data":"97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105"} Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.792466 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bda41bb8-3959-4af5-8d6f-0168d14e76f0" (UID: "bda41bb8-3959-4af5-8d6f-0168d14e76f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.804884 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xgxk8" podStartSLOduration=2.387065029 podStartE2EDuration="3.80486464s" podCreationTimestamp="2025-11-24 15:04:46 +0000 UTC" firstStartedPulling="2025-11-24 15:04:47.752100802 +0000 UTC m=+2724.868741289" lastFinishedPulling="2025-11-24 15:04:49.169900413 +0000 UTC m=+2726.286540900" observedRunningTime="2025-11-24 15:04:49.797632893 +0000 UTC m=+2726.914273410" watchObservedRunningTime="2025-11-24 15:04:49.80486464 +0000 UTC m=+2726.921505117" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.823157 4822 scope.go:117] "RemoveContainer" containerID="e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.847253 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.847292 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda41bb8-3959-4af5-8d6f-0168d14e76f0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.847307 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrgfd\" (UniqueName: \"kubernetes.io/projected/bda41bb8-3959-4af5-8d6f-0168d14e76f0-kube-api-access-vrgfd\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.853911 4822 scope.go:117] "RemoveContainer" containerID="fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.909344 4822 scope.go:117] "RemoveContainer" containerID="df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff" Nov 24 15:04:49 crc kubenswrapper[4822]: E1124 15:04:49.909798 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff\": container with ID starting with df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff not found: ID does not exist" containerID="df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.909867 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff"} err="failed to get container status \"df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff\": rpc error: code = NotFound desc = could not find container \"df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff\": container with ID starting with df66da98bfd9f252f66f1ce336ab1a51f917f757b9cdf92d0f144f6c5d6326ff not found: ID does not exist" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.909902 4822 scope.go:117] "RemoveContainer" containerID="e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba" Nov 24 15:04:49 crc kubenswrapper[4822]: E1124 15:04:49.910254 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba\": container with ID starting with e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba not found: ID does not exist" containerID="e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.910285 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba"} err="failed to get container status \"e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba\": rpc error: code = NotFound desc = could not find container \"e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba\": container with ID starting with e1d155836425d851205144dbeb4a79caca55672120780e55b7fb2f58579934ba not found: ID does not exist" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.910307 4822 scope.go:117] "RemoveContainer" containerID="fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9" Nov 24 15:04:49 crc kubenswrapper[4822]: E1124 15:04:49.910586 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9\": container with ID starting with fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9 not found: ID does not exist" containerID="fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9" Nov 24 15:04:49 crc kubenswrapper[4822]: I1124 15:04:49.910620 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9"} err="failed to get container status \"fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9\": rpc error: code = NotFound desc = could not find container \"fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9\": container with ID starting with fefd6b1ee4ec87c2a42973ca5e5935e3ee294ef3ce2a7da2c3c2a6f6c3556cd9 not found: ID does not exist" Nov 24 15:04:50 crc kubenswrapper[4822]: I1124 15:04:50.110140 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ppvxt"] Nov 24 15:04:50 crc kubenswrapper[4822]: I1124 15:04:50.119545 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ppvxt"] Nov 24 15:04:51 crc kubenswrapper[4822]: I1124 15:04:51.719270 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" path="/var/lib/kubelet/pods/bda41bb8-3959-4af5-8d6f-0168d14e76f0/volumes" Nov 24 15:04:56 crc kubenswrapper[4822]: I1124 15:04:56.462676 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:56 crc kubenswrapper[4822]: I1124 15:04:56.463066 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:56 crc kubenswrapper[4822]: I1124 15:04:56.518294 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:56 crc kubenswrapper[4822]: I1124 15:04:56.950675 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:57 crc kubenswrapper[4822]: I1124 15:04:57.010198 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgxk8"] Nov 24 15:04:58 crc kubenswrapper[4822]: I1124 15:04:58.884130 4822 generic.go:334] "Generic (PLEG): container finished" podID="f6f36194-152e-41f6-9fa9-fd0f7b279843" containerID="a5f4a2526f20d25f29c8f13ae3dc289a8633023fb543e385d3f958977e053aaa" exitCode=0 Nov 24 15:04:58 crc kubenswrapper[4822]: I1124 15:04:58.884302 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" event={"ID":"f6f36194-152e-41f6-9fa9-fd0f7b279843","Type":"ContainerDied","Data":"a5f4a2526f20d25f29c8f13ae3dc289a8633023fb543e385d3f958977e053aaa"} Nov 24 15:04:58 crc kubenswrapper[4822]: I1124 15:04:58.885259 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xgxk8" podUID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerName="registry-server" containerID="cri-o://97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105" gracePeriod=2 Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.471230 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.569025 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-catalog-content\") pod \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.569138 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-utilities\") pod \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.569180 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72k62\" (UniqueName: \"kubernetes.io/projected/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-kube-api-access-72k62\") pod \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\" (UID: \"c812e1c0-612a-43ed-a767-0da7cf5fa9e3\") " Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.570344 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-utilities" (OuterVolumeSpecName: "utilities") pod "c812e1c0-612a-43ed-a767-0da7cf5fa9e3" (UID: "c812e1c0-612a-43ed-a767-0da7cf5fa9e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.582376 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-kube-api-access-72k62" (OuterVolumeSpecName: "kube-api-access-72k62") pod "c812e1c0-612a-43ed-a767-0da7cf5fa9e3" (UID: "c812e1c0-612a-43ed-a767-0da7cf5fa9e3"). InnerVolumeSpecName "kube-api-access-72k62". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.633961 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c812e1c0-612a-43ed-a767-0da7cf5fa9e3" (UID: "c812e1c0-612a-43ed-a767-0da7cf5fa9e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.672426 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.672466 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.672481 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72k62\" (UniqueName: \"kubernetes.io/projected/c812e1c0-612a-43ed-a767-0da7cf5fa9e3-kube-api-access-72k62\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.897349 4822 generic.go:334] "Generic (PLEG): container finished" podID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerID="97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105" exitCode=0 Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.897542 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgxk8" event={"ID":"c812e1c0-612a-43ed-a767-0da7cf5fa9e3","Type":"ContainerDied","Data":"97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105"} Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.897619 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgxk8" event={"ID":"c812e1c0-612a-43ed-a767-0da7cf5fa9e3","Type":"ContainerDied","Data":"e6033072bf6c0efea3a78fbcbec8e5d7bb35c0325baff9cc04d9494bc9366eb5"} Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.897589 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgxk8" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.897696 4822 scope.go:117] "RemoveContainer" containerID="97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.929369 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgxk8"] Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.941479 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xgxk8"] Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.944438 4822 scope.go:117] "RemoveContainer" containerID="28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913" Nov 24 15:04:59 crc kubenswrapper[4822]: I1124 15:04:59.976608 4822 scope.go:117] "RemoveContainer" containerID="8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.038066 4822 scope.go:117] "RemoveContainer" containerID="97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105" Nov 24 15:05:00 crc kubenswrapper[4822]: E1124 15:05:00.038618 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105\": container with ID starting with 97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105 not found: ID does not exist" containerID="97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.038678 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105"} err="failed to get container status \"97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105\": rpc error: code = NotFound desc = could not find container \"97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105\": container with ID starting with 97dd1820fba69a6e6ae50f814b747350df10ec67c7a1c6f524c53ece29623105 not found: ID does not exist" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.038712 4822 scope.go:117] "RemoveContainer" containerID="28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913" Nov 24 15:05:00 crc kubenswrapper[4822]: E1124 15:05:00.039164 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913\": container with ID starting with 28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913 not found: ID does not exist" containerID="28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.039196 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913"} err="failed to get container status \"28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913\": rpc error: code = NotFound desc = could not find container \"28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913\": container with ID starting with 28293fa5e93715805ea6045a0db39ca9039e4dcab52f132db9f236e885c7d913 not found: ID does not exist" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.039237 4822 scope.go:117] "RemoveContainer" containerID="8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8" Nov 24 15:05:00 crc kubenswrapper[4822]: E1124 15:05:00.039585 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8\": container with ID starting with 8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8 not found: ID does not exist" containerID="8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.039622 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8"} err="failed to get container status \"8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8\": rpc error: code = NotFound desc = could not find container \"8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8\": container with ID starting with 8a015d30d2f6a274e2779580e832135526eaf765e06d4eac3455f2e776ecbbb8 not found: ID does not exist" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.482000 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.592487 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-inventory\") pod \"f6f36194-152e-41f6-9fa9-fd0f7b279843\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.592806 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-metadata-combined-ca-bundle\") pod \"f6f36194-152e-41f6-9fa9-fd0f7b279843\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.592850 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-ssh-key\") pod \"f6f36194-152e-41f6-9fa9-fd0f7b279843\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.592918 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qb7r\" (UniqueName: \"kubernetes.io/projected/f6f36194-152e-41f6-9fa9-fd0f7b279843-kube-api-access-5qb7r\") pod \"f6f36194-152e-41f6-9fa9-fd0f7b279843\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.593067 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-nova-metadata-neutron-config-0\") pod \"f6f36194-152e-41f6-9fa9-fd0f7b279843\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.593123 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-ovn-metadata-agent-neutron-config-0\") pod \"f6f36194-152e-41f6-9fa9-fd0f7b279843\" (UID: \"f6f36194-152e-41f6-9fa9-fd0f7b279843\") " Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.597583 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f36194-152e-41f6-9fa9-fd0f7b279843-kube-api-access-5qb7r" (OuterVolumeSpecName: "kube-api-access-5qb7r") pod "f6f36194-152e-41f6-9fa9-fd0f7b279843" (UID: "f6f36194-152e-41f6-9fa9-fd0f7b279843"). InnerVolumeSpecName "kube-api-access-5qb7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.599243 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f6f36194-152e-41f6-9fa9-fd0f7b279843" (UID: "f6f36194-152e-41f6-9fa9-fd0f7b279843"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.621190 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "f6f36194-152e-41f6-9fa9-fd0f7b279843" (UID: "f6f36194-152e-41f6-9fa9-fd0f7b279843"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.626576 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f6f36194-152e-41f6-9fa9-fd0f7b279843" (UID: "f6f36194-152e-41f6-9fa9-fd0f7b279843"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.628116 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "f6f36194-152e-41f6-9fa9-fd0f7b279843" (UID: "f6f36194-152e-41f6-9fa9-fd0f7b279843"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.635080 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-inventory" (OuterVolumeSpecName: "inventory") pod "f6f36194-152e-41f6-9fa9-fd0f7b279843" (UID: "f6f36194-152e-41f6-9fa9-fd0f7b279843"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.695790 4822 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.695822 4822 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.695835 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.695844 4822 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.695855 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6f36194-152e-41f6-9fa9-fd0f7b279843-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.695863 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qb7r\" (UniqueName: \"kubernetes.io/projected/f6f36194-152e-41f6-9fa9-fd0f7b279843-kube-api-access-5qb7r\") on node \"crc\" DevicePath \"\"" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.913599 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" event={"ID":"f6f36194-152e-41f6-9fa9-fd0f7b279843","Type":"ContainerDied","Data":"f86b03d4df201ea4ffadf77544259f551188f64d08c4dbbbd0746bfd77c58fa4"} Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.913646 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f86b03d4df201ea4ffadf77544259f551188f64d08c4dbbbd0746bfd77c58fa4" Nov 24 15:05:00 crc kubenswrapper[4822]: I1124 15:05:00.913710 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.056556 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn"] Nov 24 15:05:01 crc kubenswrapper[4822]: E1124 15:05:01.057103 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerName="extract-content" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057160 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerName="extract-content" Nov 24 15:05:01 crc kubenswrapper[4822]: E1124 15:05:01.057170 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerName="registry-server" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057177 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerName="registry-server" Nov 24 15:05:01 crc kubenswrapper[4822]: E1124 15:05:01.057193 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerName="registry-server" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057200 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerName="registry-server" Nov 24 15:05:01 crc kubenswrapper[4822]: E1124 15:05:01.057233 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerName="extract-utilities" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057242 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerName="extract-utilities" Nov 24 15:05:01 crc kubenswrapper[4822]: E1124 15:05:01.057269 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f36194-152e-41f6-9fa9-fd0f7b279843" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057278 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f36194-152e-41f6-9fa9-fd0f7b279843" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 15:05:01 crc kubenswrapper[4822]: E1124 15:05:01.057302 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerName="extract-content" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057310 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerName="extract-content" Nov 24 15:05:01 crc kubenswrapper[4822]: E1124 15:05:01.057325 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerName="extract-utilities" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057332 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerName="extract-utilities" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057560 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" containerName="registry-server" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057583 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="bda41bb8-3959-4af5-8d6f-0168d14e76f0" containerName="registry-server" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.057600 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f36194-152e-41f6-9fa9-fd0f7b279843" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.058596 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.061096 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.061426 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.061553 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.064787 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.064823 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.084243 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn"] Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.102557 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.102829 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.102938 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.103072 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q886z\" (UniqueName: \"kubernetes.io/projected/4b034bf8-6635-477d-a31f-b7fdf270d374-kube-api-access-q886z\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.103298 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.204957 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.205012 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.205046 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q886z\" (UniqueName: \"kubernetes.io/projected/4b034bf8-6635-477d-a31f-b7fdf270d374-kube-api-access-q886z\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.205128 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.205195 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.211021 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.211021 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.211175 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.212122 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.222771 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q886z\" (UniqueName: \"kubernetes.io/projected/4b034bf8-6635-477d-a31f-b7fdf270d374-kube-api-access-q886z\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.386329 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.725029 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c812e1c0-612a-43ed-a767-0da7cf5fa9e3" path="/var/lib/kubelet/pods/c812e1c0-612a-43ed-a767-0da7cf5fa9e3/volumes" Nov 24 15:05:01 crc kubenswrapper[4822]: I1124 15:05:01.935907 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn"] Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.167783 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cpcz8"] Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.171009 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.176703 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpcz8"] Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.228072 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-utilities\") pod \"redhat-marketplace-cpcz8\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.228264 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krp9k\" (UniqueName: \"kubernetes.io/projected/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-kube-api-access-krp9k\") pod \"redhat-marketplace-cpcz8\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.228296 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-catalog-content\") pod \"redhat-marketplace-cpcz8\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.329593 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-catalog-content\") pod \"redhat-marketplace-cpcz8\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.330121 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-utilities\") pod \"redhat-marketplace-cpcz8\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.330267 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krp9k\" (UniqueName: \"kubernetes.io/projected/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-kube-api-access-krp9k\") pod \"redhat-marketplace-cpcz8\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.330020 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-catalog-content\") pod \"redhat-marketplace-cpcz8\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.330803 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-utilities\") pod \"redhat-marketplace-cpcz8\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.350366 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krp9k\" (UniqueName: \"kubernetes.io/projected/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-kube-api-access-krp9k\") pod \"redhat-marketplace-cpcz8\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.502932 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.950055 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" event={"ID":"4b034bf8-6635-477d-a31f-b7fdf270d374","Type":"ContainerStarted","Data":"c02b3755d2c0ad4b9c116ba777c78d5e9cf3ec7ca930c052803e5346fbd4fa84"} Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.950537 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" event={"ID":"4b034bf8-6635-477d-a31f-b7fdf270d374","Type":"ContainerStarted","Data":"7e07e43e7a872856397bdf2f4c4b637c70fe26a08e5eea2f64c3a2d22a863849"} Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.959374 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpcz8"] Nov 24 15:05:02 crc kubenswrapper[4822]: W1124 15:05:02.964578 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaeb9c4f9_00c5_4899_87a8_e56efab12b9f.slice/crio-39e098149eb18e9746d60b8754a102d7752d71d988917c8cfa40046e17eb7944 WatchSource:0}: Error finding container 39e098149eb18e9746d60b8754a102d7752d71d988917c8cfa40046e17eb7944: Status 404 returned error can't find the container with id 39e098149eb18e9746d60b8754a102d7752d71d988917c8cfa40046e17eb7944 Nov 24 15:05:02 crc kubenswrapper[4822]: I1124 15:05:02.967782 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" podStartSLOduration=1.354154033 podStartE2EDuration="1.967768431s" podCreationTimestamp="2025-11-24 15:05:01 +0000 UTC" firstStartedPulling="2025-11-24 15:05:01.944471678 +0000 UTC m=+2739.061112155" lastFinishedPulling="2025-11-24 15:05:02.558086076 +0000 UTC m=+2739.674726553" observedRunningTime="2025-11-24 15:05:02.964498308 +0000 UTC m=+2740.081138785" watchObservedRunningTime="2025-11-24 15:05:02.967768431 +0000 UTC m=+2740.084408908" Nov 24 15:05:03 crc kubenswrapper[4822]: I1124 15:05:03.961572 4822 generic.go:334] "Generic (PLEG): container finished" podID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerID="472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c" exitCode=0 Nov 24 15:05:03 crc kubenswrapper[4822]: I1124 15:05:03.961640 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpcz8" event={"ID":"aeb9c4f9-00c5-4899-87a8-e56efab12b9f","Type":"ContainerDied","Data":"472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c"} Nov 24 15:05:03 crc kubenswrapper[4822]: I1124 15:05:03.961911 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpcz8" event={"ID":"aeb9c4f9-00c5-4899-87a8-e56efab12b9f","Type":"ContainerStarted","Data":"39e098149eb18e9746d60b8754a102d7752d71d988917c8cfa40046e17eb7944"} Nov 24 15:05:04 crc kubenswrapper[4822]: I1124 15:05:04.975139 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpcz8" event={"ID":"aeb9c4f9-00c5-4899-87a8-e56efab12b9f","Type":"ContainerStarted","Data":"7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167"} Nov 24 15:05:05 crc kubenswrapper[4822]: I1124 15:05:05.989098 4822 generic.go:334] "Generic (PLEG): container finished" podID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerID="7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167" exitCode=0 Nov 24 15:05:05 crc kubenswrapper[4822]: I1124 15:05:05.989190 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpcz8" event={"ID":"aeb9c4f9-00c5-4899-87a8-e56efab12b9f","Type":"ContainerDied","Data":"7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167"} Nov 24 15:05:07 crc kubenswrapper[4822]: I1124 15:05:07.030504 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpcz8" event={"ID":"aeb9c4f9-00c5-4899-87a8-e56efab12b9f","Type":"ContainerStarted","Data":"09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047"} Nov 24 15:05:07 crc kubenswrapper[4822]: I1124 15:05:07.067643 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cpcz8" podStartSLOduration=2.581188444 podStartE2EDuration="5.067616419s" podCreationTimestamp="2025-11-24 15:05:02 +0000 UTC" firstStartedPulling="2025-11-24 15:05:03.965258804 +0000 UTC m=+2741.081899281" lastFinishedPulling="2025-11-24 15:05:06.451686739 +0000 UTC m=+2743.568327256" observedRunningTime="2025-11-24 15:05:07.053669861 +0000 UTC m=+2744.170310388" watchObservedRunningTime="2025-11-24 15:05:07.067616419 +0000 UTC m=+2744.184256906" Nov 24 15:05:11 crc kubenswrapper[4822]: I1124 15:05:11.592605 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:05:11 crc kubenswrapper[4822]: I1124 15:05:11.595059 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:05:11 crc kubenswrapper[4822]: I1124 15:05:11.600701 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 15:05:11 crc kubenswrapper[4822]: I1124 15:05:11.612510 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"883d279d8005f9e92a76f4480790753dc5091bbb3ea64a7c08767b819a41a4f8"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:05:11 crc kubenswrapper[4822]: I1124 15:05:11.612625 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://883d279d8005f9e92a76f4480790753dc5091bbb3ea64a7c08767b819a41a4f8" gracePeriod=600 Nov 24 15:05:12 crc kubenswrapper[4822]: I1124 15:05:12.096563 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="883d279d8005f9e92a76f4480790753dc5091bbb3ea64a7c08767b819a41a4f8" exitCode=0 Nov 24 15:05:12 crc kubenswrapper[4822]: I1124 15:05:12.096672 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"883d279d8005f9e92a76f4480790753dc5091bbb3ea64a7c08767b819a41a4f8"} Nov 24 15:05:12 crc kubenswrapper[4822]: I1124 15:05:12.097167 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102"} Nov 24 15:05:12 crc kubenswrapper[4822]: I1124 15:05:12.097227 4822 scope.go:117] "RemoveContainer" containerID="cb763fe2653cef7b7d30032c92df7f26c0e5af5bd1594cecf93530d78ed5b300" Nov 24 15:05:12 crc kubenswrapper[4822]: I1124 15:05:12.503937 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:12 crc kubenswrapper[4822]: I1124 15:05:12.504299 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:12 crc kubenswrapper[4822]: I1124 15:05:12.568773 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:13 crc kubenswrapper[4822]: I1124 15:05:13.180394 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:13 crc kubenswrapper[4822]: I1124 15:05:13.234286 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpcz8"] Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.136108 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cpcz8" podUID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerName="registry-server" containerID="cri-o://09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047" gracePeriod=2 Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.700356 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.866061 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-utilities\") pod \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.866723 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-catalog-content\") pod \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.866992 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krp9k\" (UniqueName: \"kubernetes.io/projected/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-kube-api-access-krp9k\") pod \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\" (UID: \"aeb9c4f9-00c5-4899-87a8-e56efab12b9f\") " Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.867117 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-utilities" (OuterVolumeSpecName: "utilities") pod "aeb9c4f9-00c5-4899-87a8-e56efab12b9f" (UID: "aeb9c4f9-00c5-4899-87a8-e56efab12b9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.870041 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.892702 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-kube-api-access-krp9k" (OuterVolumeSpecName: "kube-api-access-krp9k") pod "aeb9c4f9-00c5-4899-87a8-e56efab12b9f" (UID: "aeb9c4f9-00c5-4899-87a8-e56efab12b9f"). InnerVolumeSpecName "kube-api-access-krp9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.906367 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aeb9c4f9-00c5-4899-87a8-e56efab12b9f" (UID: "aeb9c4f9-00c5-4899-87a8-e56efab12b9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.973040 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:05:15 crc kubenswrapper[4822]: I1124 15:05:15.973101 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krp9k\" (UniqueName: \"kubernetes.io/projected/aeb9c4f9-00c5-4899-87a8-e56efab12b9f-kube-api-access-krp9k\") on node \"crc\" DevicePath \"\"" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.148928 4822 generic.go:334] "Generic (PLEG): container finished" podID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerID="09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047" exitCode=0 Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.148976 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpcz8" event={"ID":"aeb9c4f9-00c5-4899-87a8-e56efab12b9f","Type":"ContainerDied","Data":"09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047"} Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.149007 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpcz8" event={"ID":"aeb9c4f9-00c5-4899-87a8-e56efab12b9f","Type":"ContainerDied","Data":"39e098149eb18e9746d60b8754a102d7752d71d988917c8cfa40046e17eb7944"} Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.149027 4822 scope.go:117] "RemoveContainer" containerID="09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.149237 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cpcz8" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.192016 4822 scope.go:117] "RemoveContainer" containerID="7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.196247 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpcz8"] Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.204882 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpcz8"] Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.216681 4822 scope.go:117] "RemoveContainer" containerID="472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.288192 4822 scope.go:117] "RemoveContainer" containerID="09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047" Nov 24 15:05:16 crc kubenswrapper[4822]: E1124 15:05:16.289465 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047\": container with ID starting with 09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047 not found: ID does not exist" containerID="09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.289529 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047"} err="failed to get container status \"09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047\": rpc error: code = NotFound desc = could not find container \"09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047\": container with ID starting with 09f91a6de937a44f912de708c0e4ac4d8cae26c2b78702206ee4b07bc23d8047 not found: ID does not exist" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.289571 4822 scope.go:117] "RemoveContainer" containerID="7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167" Nov 24 15:05:16 crc kubenswrapper[4822]: E1124 15:05:16.290304 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167\": container with ID starting with 7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167 not found: ID does not exist" containerID="7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.290424 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167"} err="failed to get container status \"7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167\": rpc error: code = NotFound desc = could not find container \"7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167\": container with ID starting with 7f0aa37513264a154232db7438c1aef1507abb474b13ce7f0b3cbe520481f167 not found: ID does not exist" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.290491 4822 scope.go:117] "RemoveContainer" containerID="472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c" Nov 24 15:05:16 crc kubenswrapper[4822]: E1124 15:05:16.291129 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c\": container with ID starting with 472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c not found: ID does not exist" containerID="472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c" Nov 24 15:05:16 crc kubenswrapper[4822]: I1124 15:05:16.291162 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c"} err="failed to get container status \"472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c\": rpc error: code = NotFound desc = could not find container \"472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c\": container with ID starting with 472a1a08291e830d06d09ad7e3a5b40bbee9a6d09fde8ccd90af87681ab9a83c not found: ID does not exist" Nov 24 15:05:17 crc kubenswrapper[4822]: I1124 15:05:17.720272 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" path="/var/lib/kubelet/pods/aeb9c4f9-00c5-4899-87a8-e56efab12b9f/volumes" Nov 24 15:07:11 crc kubenswrapper[4822]: I1124 15:07:11.578352 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:07:11 crc kubenswrapper[4822]: I1124 15:07:11.579014 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.772405 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wnk5r"] Nov 24 15:07:27 crc kubenswrapper[4822]: E1124 15:07:27.773561 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerName="extract-utilities" Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.773583 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerName="extract-utilities" Nov 24 15:07:27 crc kubenswrapper[4822]: E1124 15:07:27.773641 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerName="registry-server" Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.773651 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerName="registry-server" Nov 24 15:07:27 crc kubenswrapper[4822]: E1124 15:07:27.773683 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerName="extract-content" Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.773694 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerName="extract-content" Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.773998 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb9c4f9-00c5-4899-87a8-e56efab12b9f" containerName="registry-server" Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.776542 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.784513 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wnk5r"] Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.951119 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-catalog-content\") pod \"redhat-operators-wnk5r\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.951286 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwq9r\" (UniqueName: \"kubernetes.io/projected/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-kube-api-access-zwq9r\") pod \"redhat-operators-wnk5r\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:27 crc kubenswrapper[4822]: I1124 15:07:27.951349 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-utilities\") pod \"redhat-operators-wnk5r\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:28 crc kubenswrapper[4822]: I1124 15:07:28.054091 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-catalog-content\") pod \"redhat-operators-wnk5r\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:28 crc kubenswrapper[4822]: I1124 15:07:28.054494 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwq9r\" (UniqueName: \"kubernetes.io/projected/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-kube-api-access-zwq9r\") pod \"redhat-operators-wnk5r\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:28 crc kubenswrapper[4822]: I1124 15:07:28.054593 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-catalog-content\") pod \"redhat-operators-wnk5r\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:28 crc kubenswrapper[4822]: I1124 15:07:28.054722 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-utilities\") pod \"redhat-operators-wnk5r\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:28 crc kubenswrapper[4822]: I1124 15:07:28.055285 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-utilities\") pod \"redhat-operators-wnk5r\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:28 crc kubenswrapper[4822]: I1124 15:07:28.072261 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwq9r\" (UniqueName: \"kubernetes.io/projected/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-kube-api-access-zwq9r\") pod \"redhat-operators-wnk5r\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:28 crc kubenswrapper[4822]: I1124 15:07:28.097305 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:28 crc kubenswrapper[4822]: I1124 15:07:28.593244 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wnk5r"] Nov 24 15:07:28 crc kubenswrapper[4822]: W1124 15:07:28.600016 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee618867_76dd_4aab_a5b5_b49a5a18bbe3.slice/crio-010e46dee992609197677e837878837c091c72e8eef0e87b9fcfa611aad200c4 WatchSource:0}: Error finding container 010e46dee992609197677e837878837c091c72e8eef0e87b9fcfa611aad200c4: Status 404 returned error can't find the container with id 010e46dee992609197677e837878837c091c72e8eef0e87b9fcfa611aad200c4 Nov 24 15:07:28 crc kubenswrapper[4822]: I1124 15:07:28.764342 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wnk5r" event={"ID":"ee618867-76dd-4aab-a5b5-b49a5a18bbe3","Type":"ContainerStarted","Data":"010e46dee992609197677e837878837c091c72e8eef0e87b9fcfa611aad200c4"} Nov 24 15:07:29 crc kubenswrapper[4822]: I1124 15:07:29.775606 4822 generic.go:334] "Generic (PLEG): container finished" podID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerID="5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3" exitCode=0 Nov 24 15:07:29 crc kubenswrapper[4822]: I1124 15:07:29.775669 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wnk5r" event={"ID":"ee618867-76dd-4aab-a5b5-b49a5a18bbe3","Type":"ContainerDied","Data":"5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3"} Nov 24 15:07:31 crc kubenswrapper[4822]: I1124 15:07:31.802974 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wnk5r" event={"ID":"ee618867-76dd-4aab-a5b5-b49a5a18bbe3","Type":"ContainerStarted","Data":"10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a"} Nov 24 15:07:34 crc kubenswrapper[4822]: I1124 15:07:34.851826 4822 generic.go:334] "Generic (PLEG): container finished" podID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerID="10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a" exitCode=0 Nov 24 15:07:34 crc kubenswrapper[4822]: I1124 15:07:34.851897 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wnk5r" event={"ID":"ee618867-76dd-4aab-a5b5-b49a5a18bbe3","Type":"ContainerDied","Data":"10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a"} Nov 24 15:07:35 crc kubenswrapper[4822]: I1124 15:07:35.863815 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wnk5r" event={"ID":"ee618867-76dd-4aab-a5b5-b49a5a18bbe3","Type":"ContainerStarted","Data":"80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80"} Nov 24 15:07:35 crc kubenswrapper[4822]: I1124 15:07:35.896991 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wnk5r" podStartSLOduration=3.45106171 podStartE2EDuration="8.896964615s" podCreationTimestamp="2025-11-24 15:07:27 +0000 UTC" firstStartedPulling="2025-11-24 15:07:29.778001165 +0000 UTC m=+2886.894641642" lastFinishedPulling="2025-11-24 15:07:35.22390403 +0000 UTC m=+2892.340544547" observedRunningTime="2025-11-24 15:07:35.882945236 +0000 UTC m=+2892.999585773" watchObservedRunningTime="2025-11-24 15:07:35.896964615 +0000 UTC m=+2893.013605092" Nov 24 15:07:38 crc kubenswrapper[4822]: I1124 15:07:38.098071 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:38 crc kubenswrapper[4822]: I1124 15:07:38.099029 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:39 crc kubenswrapper[4822]: I1124 15:07:39.162096 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wnk5r" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerName="registry-server" probeResult="failure" output=< Nov 24 15:07:39 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 15:07:39 crc kubenswrapper[4822]: > Nov 24 15:07:41 crc kubenswrapper[4822]: I1124 15:07:41.578475 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:07:41 crc kubenswrapper[4822]: I1124 15:07:41.578840 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:07:48 crc kubenswrapper[4822]: I1124 15:07:48.204444 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:48 crc kubenswrapper[4822]: I1124 15:07:48.277834 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:48 crc kubenswrapper[4822]: I1124 15:07:48.459455 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wnk5r"] Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.038496 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wnk5r" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerName="registry-server" containerID="cri-o://80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80" gracePeriod=2 Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.567562 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.605275 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-catalog-content\") pod \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.605378 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwq9r\" (UniqueName: \"kubernetes.io/projected/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-kube-api-access-zwq9r\") pod \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.605572 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-utilities\") pod \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\" (UID: \"ee618867-76dd-4aab-a5b5-b49a5a18bbe3\") " Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.606987 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-utilities" (OuterVolumeSpecName: "utilities") pod "ee618867-76dd-4aab-a5b5-b49a5a18bbe3" (UID: "ee618867-76dd-4aab-a5b5-b49a5a18bbe3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.614577 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-kube-api-access-zwq9r" (OuterVolumeSpecName: "kube-api-access-zwq9r") pod "ee618867-76dd-4aab-a5b5-b49a5a18bbe3" (UID: "ee618867-76dd-4aab-a5b5-b49a5a18bbe3"). InnerVolumeSpecName "kube-api-access-zwq9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.707766 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwq9r\" (UniqueName: \"kubernetes.io/projected/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-kube-api-access-zwq9r\") on node \"crc\" DevicePath \"\"" Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.707800 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.756846 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee618867-76dd-4aab-a5b5-b49a5a18bbe3" (UID: "ee618867-76dd-4aab-a5b5-b49a5a18bbe3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:07:50 crc kubenswrapper[4822]: I1124 15:07:50.809798 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee618867-76dd-4aab-a5b5-b49a5a18bbe3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.049128 4822 generic.go:334] "Generic (PLEG): container finished" podID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerID="80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80" exitCode=0 Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.049396 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wnk5r" event={"ID":"ee618867-76dd-4aab-a5b5-b49a5a18bbe3","Type":"ContainerDied","Data":"80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80"} Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.049423 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wnk5r" event={"ID":"ee618867-76dd-4aab-a5b5-b49a5a18bbe3","Type":"ContainerDied","Data":"010e46dee992609197677e837878837c091c72e8eef0e87b9fcfa611aad200c4"} Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.049440 4822 scope.go:117] "RemoveContainer" containerID="80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.049563 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wnk5r" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.087348 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wnk5r"] Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.088282 4822 scope.go:117] "RemoveContainer" containerID="10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.097144 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wnk5r"] Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.110425 4822 scope.go:117] "RemoveContainer" containerID="5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.200180 4822 scope.go:117] "RemoveContainer" containerID="80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80" Nov 24 15:07:51 crc kubenswrapper[4822]: E1124 15:07:51.200731 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80\": container with ID starting with 80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80 not found: ID does not exist" containerID="80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.200787 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80"} err="failed to get container status \"80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80\": rpc error: code = NotFound desc = could not find container \"80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80\": container with ID starting with 80f90a84c6df245bc4d90a949db47cde6b38d09d6b628f9cf91907d56456ee80 not found: ID does not exist" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.200821 4822 scope.go:117] "RemoveContainer" containerID="10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a" Nov 24 15:07:51 crc kubenswrapper[4822]: E1124 15:07:51.201418 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a\": container with ID starting with 10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a not found: ID does not exist" containerID="10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.201490 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a"} err="failed to get container status \"10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a\": rpc error: code = NotFound desc = could not find container \"10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a\": container with ID starting with 10e6427b3f6a35cfdb9e501a21386bcd88cf832db452014db1a09a9bb1603c1a not found: ID does not exist" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.201533 4822 scope.go:117] "RemoveContainer" containerID="5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3" Nov 24 15:07:51 crc kubenswrapper[4822]: E1124 15:07:51.201857 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3\": container with ID starting with 5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3 not found: ID does not exist" containerID="5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.201899 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3"} err="failed to get container status \"5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3\": rpc error: code = NotFound desc = could not find container \"5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3\": container with ID starting with 5938b2caf9b9821778027884699ed6b1da77e156a9dd09052eb1380131c53aa3 not found: ID does not exist" Nov 24 15:07:51 crc kubenswrapper[4822]: I1124 15:07:51.725563 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" path="/var/lib/kubelet/pods/ee618867-76dd-4aab-a5b5-b49a5a18bbe3/volumes" Nov 24 15:08:11 crc kubenswrapper[4822]: I1124 15:08:11.578003 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:08:11 crc kubenswrapper[4822]: I1124 15:08:11.578601 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:08:11 crc kubenswrapper[4822]: I1124 15:08:11.578653 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 15:08:11 crc kubenswrapper[4822]: I1124 15:08:11.579457 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:08:11 crc kubenswrapper[4822]: I1124 15:08:11.579519 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" gracePeriod=600 Nov 24 15:08:11 crc kubenswrapper[4822]: E1124 15:08:11.700997 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:08:12 crc kubenswrapper[4822]: I1124 15:08:12.307845 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" exitCode=0 Nov 24 15:08:12 crc kubenswrapper[4822]: I1124 15:08:12.307942 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102"} Nov 24 15:08:12 crc kubenswrapper[4822]: I1124 15:08:12.308186 4822 scope.go:117] "RemoveContainer" containerID="883d279d8005f9e92a76f4480790753dc5091bbb3ea64a7c08767b819a41a4f8" Nov 24 15:08:12 crc kubenswrapper[4822]: I1124 15:08:12.309353 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:08:12 crc kubenswrapper[4822]: E1124 15:08:12.313333 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:08:24 crc kubenswrapper[4822]: I1124 15:08:24.705270 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:08:24 crc kubenswrapper[4822]: E1124 15:08:24.705982 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:08:38 crc kubenswrapper[4822]: I1124 15:08:38.704468 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:08:38 crc kubenswrapper[4822]: E1124 15:08:38.705077 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:08:50 crc kubenswrapper[4822]: I1124 15:08:50.705169 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:08:50 crc kubenswrapper[4822]: E1124 15:08:50.707050 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:09:03 crc kubenswrapper[4822]: I1124 15:09:03.710107 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:09:03 crc kubenswrapper[4822]: E1124 15:09:03.710899 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:09:17 crc kubenswrapper[4822]: I1124 15:09:17.705030 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:09:17 crc kubenswrapper[4822]: E1124 15:09:17.705750 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:09:30 crc kubenswrapper[4822]: I1124 15:09:30.705759 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:09:30 crc kubenswrapper[4822]: E1124 15:09:30.706847 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:09:39 crc kubenswrapper[4822]: I1124 15:09:39.293485 4822 generic.go:334] "Generic (PLEG): container finished" podID="4b034bf8-6635-477d-a31f-b7fdf270d374" containerID="c02b3755d2c0ad4b9c116ba777c78d5e9cf3ec7ca930c052803e5346fbd4fa84" exitCode=0 Nov 24 15:09:39 crc kubenswrapper[4822]: I1124 15:09:39.293640 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" event={"ID":"4b034bf8-6635-477d-a31f-b7fdf270d374","Type":"ContainerDied","Data":"c02b3755d2c0ad4b9c116ba777c78d5e9cf3ec7ca930c052803e5346fbd4fa84"} Nov 24 15:09:40 crc kubenswrapper[4822]: I1124 15:09:40.830961 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:09:40 crc kubenswrapper[4822]: I1124 15:09:40.963602 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-combined-ca-bundle\") pod \"4b034bf8-6635-477d-a31f-b7fdf270d374\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " Nov 24 15:09:40 crc kubenswrapper[4822]: I1124 15:09:40.963667 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q886z\" (UniqueName: \"kubernetes.io/projected/4b034bf8-6635-477d-a31f-b7fdf270d374-kube-api-access-q886z\") pod \"4b034bf8-6635-477d-a31f-b7fdf270d374\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " Nov 24 15:09:40 crc kubenswrapper[4822]: I1124 15:09:40.963791 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-ssh-key\") pod \"4b034bf8-6635-477d-a31f-b7fdf270d374\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " Nov 24 15:09:40 crc kubenswrapper[4822]: I1124 15:09:40.963976 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-secret-0\") pod \"4b034bf8-6635-477d-a31f-b7fdf270d374\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " Nov 24 15:09:40 crc kubenswrapper[4822]: I1124 15:09:40.964015 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-inventory\") pod \"4b034bf8-6635-477d-a31f-b7fdf270d374\" (UID: \"4b034bf8-6635-477d-a31f-b7fdf270d374\") " Nov 24 15:09:40 crc kubenswrapper[4822]: I1124 15:09:40.969337 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4b034bf8-6635-477d-a31f-b7fdf270d374" (UID: "4b034bf8-6635-477d-a31f-b7fdf270d374"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:09:40 crc kubenswrapper[4822]: I1124 15:09:40.970190 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b034bf8-6635-477d-a31f-b7fdf270d374-kube-api-access-q886z" (OuterVolumeSpecName: "kube-api-access-q886z") pod "4b034bf8-6635-477d-a31f-b7fdf270d374" (UID: "4b034bf8-6635-477d-a31f-b7fdf270d374"). InnerVolumeSpecName "kube-api-access-q886z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:09:40 crc kubenswrapper[4822]: I1124 15:09:40.999828 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-inventory" (OuterVolumeSpecName: "inventory") pod "4b034bf8-6635-477d-a31f-b7fdf270d374" (UID: "4b034bf8-6635-477d-a31f-b7fdf270d374"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.003390 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b034bf8-6635-477d-a31f-b7fdf270d374" (UID: "4b034bf8-6635-477d-a31f-b7fdf270d374"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.005046 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4b034bf8-6635-477d-a31f-b7fdf270d374" (UID: "4b034bf8-6635-477d-a31f-b7fdf270d374"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.067060 4822 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.067090 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.067100 4822 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.067110 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q886z\" (UniqueName: \"kubernetes.io/projected/4b034bf8-6635-477d-a31f-b7fdf270d374-kube-api-access-q886z\") on node \"crc\" DevicePath \"\"" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.067120 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b034bf8-6635-477d-a31f-b7fdf270d374-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.312572 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" event={"ID":"4b034bf8-6635-477d-a31f-b7fdf270d374","Type":"ContainerDied","Data":"7e07e43e7a872856397bdf2f4c4b637c70fe26a08e5eea2f64c3a2d22a863849"} Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.312629 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e07e43e7a872856397bdf2f4c4b637c70fe26a08e5eea2f64c3a2d22a863849" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.312668 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.447230 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9"] Nov 24 15:09:41 crc kubenswrapper[4822]: E1124 15:09:41.447719 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerName="extract-content" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.447736 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerName="extract-content" Nov 24 15:09:41 crc kubenswrapper[4822]: E1124 15:09:41.447751 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b034bf8-6635-477d-a31f-b7fdf270d374" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.447761 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b034bf8-6635-477d-a31f-b7fdf270d374" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 15:09:41 crc kubenswrapper[4822]: E1124 15:09:41.447785 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerName="extract-utilities" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.447793 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerName="extract-utilities" Nov 24 15:09:41 crc kubenswrapper[4822]: E1124 15:09:41.447833 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerName="registry-server" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.447842 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerName="registry-server" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.448118 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee618867-76dd-4aab-a5b5-b49a5a18bbe3" containerName="registry-server" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.448139 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b034bf8-6635-477d-a31f-b7fdf270d374" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.449101 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.452033 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.452071 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.452543 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.453438 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.454252 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.458771 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.461885 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.469242 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9"] Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.577659 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.577701 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.577722 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.577766 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.577794 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.577900 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.577937 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.577981 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dqmf\" (UniqueName: \"kubernetes.io/projected/cb850024-1cff-4a61-8d54-c0eff343bd4b-kube-api-access-7dqmf\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.578295 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.680978 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.681119 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.681178 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.681240 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.681331 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.681396 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.681513 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.681561 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.681656 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dqmf\" (UniqueName: \"kubernetes.io/projected/cb850024-1cff-4a61-8d54-c0eff343bd4b-kube-api-access-7dqmf\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.682635 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.685673 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.685957 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.686259 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.686799 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.688054 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.688510 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.689633 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.711570 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dqmf\" (UniqueName: \"kubernetes.io/projected/cb850024-1cff-4a61-8d54-c0eff343bd4b-kube-api-access-7dqmf\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cfwv9\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:41 crc kubenswrapper[4822]: I1124 15:09:41.799063 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:09:42 crc kubenswrapper[4822]: I1124 15:09:42.461354 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9"] Nov 24 15:09:42 crc kubenswrapper[4822]: I1124 15:09:42.473981 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:09:42 crc kubenswrapper[4822]: I1124 15:09:42.704898 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:09:42 crc kubenswrapper[4822]: E1124 15:09:42.705244 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:09:43 crc kubenswrapper[4822]: I1124 15:09:43.342881 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" event={"ID":"cb850024-1cff-4a61-8d54-c0eff343bd4b","Type":"ContainerStarted","Data":"cabecb2e7fb261e6bc405347262a7ccd45b3225ef4d31b05dc767fcbdc2bb34a"} Nov 24 15:09:43 crc kubenswrapper[4822]: I1124 15:09:43.343408 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" event={"ID":"cb850024-1cff-4a61-8d54-c0eff343bd4b","Type":"ContainerStarted","Data":"f3c70ccfe3f081a8df1e0372f954cc54e569997830d0ccf8e9c978238f544201"} Nov 24 15:09:43 crc kubenswrapper[4822]: I1124 15:09:43.379773 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" podStartSLOduration=1.901120371 podStartE2EDuration="2.379739343s" podCreationTimestamp="2025-11-24 15:09:41 +0000 UTC" firstStartedPulling="2025-11-24 15:09:42.473515566 +0000 UTC m=+3019.590156043" lastFinishedPulling="2025-11-24 15:09:42.952134508 +0000 UTC m=+3020.068775015" observedRunningTime="2025-11-24 15:09:43.368637855 +0000 UTC m=+3020.485278362" watchObservedRunningTime="2025-11-24 15:09:43.379739343 +0000 UTC m=+3020.496379820" Nov 24 15:09:56 crc kubenswrapper[4822]: I1124 15:09:56.704826 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:09:56 crc kubenswrapper[4822]: E1124 15:09:56.705876 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:10:09 crc kubenswrapper[4822]: I1124 15:10:09.708183 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:10:09 crc kubenswrapper[4822]: E1124 15:10:09.709128 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:10:20 crc kubenswrapper[4822]: I1124 15:10:20.705084 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:10:20 crc kubenswrapper[4822]: E1124 15:10:20.706466 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:10:34 crc kubenswrapper[4822]: I1124 15:10:34.704525 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:10:34 crc kubenswrapper[4822]: E1124 15:10:34.705627 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:10:49 crc kubenswrapper[4822]: I1124 15:10:49.705800 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:10:49 crc kubenswrapper[4822]: E1124 15:10:49.706630 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:11:01 crc kubenswrapper[4822]: I1124 15:11:01.705609 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:11:01 crc kubenswrapper[4822]: E1124 15:11:01.706466 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:11:15 crc kubenswrapper[4822]: I1124 15:11:15.704801 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:11:15 crc kubenswrapper[4822]: E1124 15:11:15.705399 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:11:26 crc kubenswrapper[4822]: I1124 15:11:26.705470 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:11:26 crc kubenswrapper[4822]: E1124 15:11:26.706232 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:11:37 crc kubenswrapper[4822]: I1124 15:11:37.705651 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:11:37 crc kubenswrapper[4822]: E1124 15:11:37.706542 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:11:50 crc kubenswrapper[4822]: I1124 15:11:50.705380 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:11:50 crc kubenswrapper[4822]: E1124 15:11:50.706606 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:12:01 crc kubenswrapper[4822]: I1124 15:12:01.704700 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:12:01 crc kubenswrapper[4822]: E1124 15:12:01.705859 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:12:13 crc kubenswrapper[4822]: I1124 15:12:13.711859 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:12:13 crc kubenswrapper[4822]: E1124 15:12:13.712645 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:12:27 crc kubenswrapper[4822]: I1124 15:12:27.704821 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:12:27 crc kubenswrapper[4822]: E1124 15:12:27.705787 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:12:38 crc kubenswrapper[4822]: I1124 15:12:38.705304 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:12:38 crc kubenswrapper[4822]: E1124 15:12:38.706496 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:12:51 crc kubenswrapper[4822]: I1124 15:12:51.705429 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:12:51 crc kubenswrapper[4822]: E1124 15:12:51.706829 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:12:52 crc kubenswrapper[4822]: I1124 15:12:52.699136 4822 generic.go:334] "Generic (PLEG): container finished" podID="cb850024-1cff-4a61-8d54-c0eff343bd4b" containerID="cabecb2e7fb261e6bc405347262a7ccd45b3225ef4d31b05dc767fcbdc2bb34a" exitCode=0 Nov 24 15:12:52 crc kubenswrapper[4822]: I1124 15:12:52.699260 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" event={"ID":"cb850024-1cff-4a61-8d54-c0eff343bd4b","Type":"ContainerDied","Data":"cabecb2e7fb261e6bc405347262a7ccd45b3225ef4d31b05dc767fcbdc2bb34a"} Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.333653 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.502923 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dqmf\" (UniqueName: \"kubernetes.io/projected/cb850024-1cff-4a61-8d54-c0eff343bd4b-kube-api-access-7dqmf\") pod \"cb850024-1cff-4a61-8d54-c0eff343bd4b\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.502971 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-ssh-key\") pod \"cb850024-1cff-4a61-8d54-c0eff343bd4b\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.502993 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-0\") pod \"cb850024-1cff-4a61-8d54-c0eff343bd4b\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.503891 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-combined-ca-bundle\") pod \"cb850024-1cff-4a61-8d54-c0eff343bd4b\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.503929 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-0\") pod \"cb850024-1cff-4a61-8d54-c0eff343bd4b\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.503965 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-1\") pod \"cb850024-1cff-4a61-8d54-c0eff343bd4b\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.504047 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-1\") pod \"cb850024-1cff-4a61-8d54-c0eff343bd4b\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.504078 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-extra-config-0\") pod \"cb850024-1cff-4a61-8d54-c0eff343bd4b\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.504159 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-inventory\") pod \"cb850024-1cff-4a61-8d54-c0eff343bd4b\" (UID: \"cb850024-1cff-4a61-8d54-c0eff343bd4b\") " Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.521417 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb850024-1cff-4a61-8d54-c0eff343bd4b-kube-api-access-7dqmf" (OuterVolumeSpecName: "kube-api-access-7dqmf") pod "cb850024-1cff-4a61-8d54-c0eff343bd4b" (UID: "cb850024-1cff-4a61-8d54-c0eff343bd4b"). InnerVolumeSpecName "kube-api-access-7dqmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.523320 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "cb850024-1cff-4a61-8d54-c0eff343bd4b" (UID: "cb850024-1cff-4a61-8d54-c0eff343bd4b"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.531526 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cb850024-1cff-4a61-8d54-c0eff343bd4b" (UID: "cb850024-1cff-4a61-8d54-c0eff343bd4b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.532909 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-inventory" (OuterVolumeSpecName: "inventory") pod "cb850024-1cff-4a61-8d54-c0eff343bd4b" (UID: "cb850024-1cff-4a61-8d54-c0eff343bd4b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.535609 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "cb850024-1cff-4a61-8d54-c0eff343bd4b" (UID: "cb850024-1cff-4a61-8d54-c0eff343bd4b"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.541913 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "cb850024-1cff-4a61-8d54-c0eff343bd4b" (UID: "cb850024-1cff-4a61-8d54-c0eff343bd4b"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.543358 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "cb850024-1cff-4a61-8d54-c0eff343bd4b" (UID: "cb850024-1cff-4a61-8d54-c0eff343bd4b"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.551856 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "cb850024-1cff-4a61-8d54-c0eff343bd4b" (UID: "cb850024-1cff-4a61-8d54-c0eff343bd4b"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.559744 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "cb850024-1cff-4a61-8d54-c0eff343bd4b" (UID: "cb850024-1cff-4a61-8d54-c0eff343bd4b"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.606364 4822 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.606394 4822 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.606405 4822 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.606413 4822 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.606422 4822 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.606431 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.606441 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dqmf\" (UniqueName: \"kubernetes.io/projected/cb850024-1cff-4a61-8d54-c0eff343bd4b-kube-api-access-7dqmf\") on node \"crc\" DevicePath \"\"" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.606449 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.606457 4822 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb850024-1cff-4a61-8d54-c0eff343bd4b-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.722073 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" event={"ID":"cb850024-1cff-4a61-8d54-c0eff343bd4b","Type":"ContainerDied","Data":"f3c70ccfe3f081a8df1e0372f954cc54e569997830d0ccf8e9c978238f544201"} Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.722109 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3c70ccfe3f081a8df1e0372f954cc54e569997830d0ccf8e9c978238f544201" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.722149 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cfwv9" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.833293 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg"] Nov 24 15:12:54 crc kubenswrapper[4822]: E1124 15:12:54.833830 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb850024-1cff-4a61-8d54-c0eff343bd4b" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.833850 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb850024-1cff-4a61-8d54-c0eff343bd4b" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.834172 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb850024-1cff-4a61-8d54-c0eff343bd4b" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.835832 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.837819 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.839609 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.839781 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.839906 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9bjhw" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.839988 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 15:12:54 crc kubenswrapper[4822]: I1124 15:12:54.848859 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg"] Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.014394 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.014495 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.014554 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.014594 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.014686 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.014771 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bthgd\" (UniqueName: \"kubernetes.io/projected/561e27e7-d8ea-4764-872e-0b022659275c-kube-api-access-bthgd\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.014907 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.117051 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.117193 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.117326 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.117390 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.117525 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.117657 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bthgd\" (UniqueName: \"kubernetes.io/projected/561e27e7-d8ea-4764-872e-0b022659275c-kube-api-access-bthgd\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.117875 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.122078 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.123497 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.134646 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.134793 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.136022 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.139596 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.149605 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bthgd\" (UniqueName: \"kubernetes.io/projected/561e27e7-d8ea-4764-872e-0b022659275c-kube-api-access-bthgd\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s22zg\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.166782 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:12:55 crc kubenswrapper[4822]: I1124 15:12:55.824363 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg"] Nov 24 15:12:56 crc kubenswrapper[4822]: I1124 15:12:56.752909 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" event={"ID":"561e27e7-d8ea-4764-872e-0b022659275c","Type":"ContainerStarted","Data":"34d88b00b7f3fe35002b6add35967d052817b15de64c99d0cf68ae185094ae90"} Nov 24 15:12:56 crc kubenswrapper[4822]: I1124 15:12:56.753299 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" event={"ID":"561e27e7-d8ea-4764-872e-0b022659275c","Type":"ContainerStarted","Data":"a0a1d2bc7c3f588486fa3b49034234f2700f9e8cb1ded4887568607420a03886"} Nov 24 15:12:56 crc kubenswrapper[4822]: I1124 15:12:56.791526 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" podStartSLOduration=2.268678356 podStartE2EDuration="2.791497504s" podCreationTimestamp="2025-11-24 15:12:54 +0000 UTC" firstStartedPulling="2025-11-24 15:12:55.827761093 +0000 UTC m=+3212.944401590" lastFinishedPulling="2025-11-24 15:12:56.350580261 +0000 UTC m=+3213.467220738" observedRunningTime="2025-11-24 15:12:56.776079952 +0000 UTC m=+3213.892720459" watchObservedRunningTime="2025-11-24 15:12:56.791497504 +0000 UTC m=+3213.908138031" Nov 24 15:13:05 crc kubenswrapper[4822]: I1124 15:13:05.704616 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:13:05 crc kubenswrapper[4822]: E1124 15:13:05.705800 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:13:17 crc kubenswrapper[4822]: I1124 15:13:17.704681 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:13:18 crc kubenswrapper[4822]: I1124 15:13:18.007472 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"7b89a251fd0f0cfe8d88737b003ae9166ecdc96e5233f626b16b042314ed274c"} Nov 24 15:14:49 crc kubenswrapper[4822]: I1124 15:14:49.877824 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k4tjc"] Nov 24 15:14:49 crc kubenswrapper[4822]: I1124 15:14:49.882779 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:49 crc kubenswrapper[4822]: I1124 15:14:49.897931 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k4tjc"] Nov 24 15:14:49 crc kubenswrapper[4822]: I1124 15:14:49.965645 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvfvv\" (UniqueName: \"kubernetes.io/projected/189cb4b9-5c76-4872-b350-29525973b831-kube-api-access-jvfvv\") pod \"certified-operators-k4tjc\" (UID: \"189cb4b9-5c76-4872-b350-29525973b831\") " pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:49 crc kubenswrapper[4822]: I1124 15:14:49.965724 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/189cb4b9-5c76-4872-b350-29525973b831-utilities\") pod \"certified-operators-k4tjc\" (UID: \"189cb4b9-5c76-4872-b350-29525973b831\") " pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:49 crc kubenswrapper[4822]: I1124 15:14:49.966134 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/189cb4b9-5c76-4872-b350-29525973b831-catalog-content\") pod \"certified-operators-k4tjc\" (UID: \"189cb4b9-5c76-4872-b350-29525973b831\") " pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:50 crc kubenswrapper[4822]: I1124 15:14:50.069005 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvfvv\" (UniqueName: \"kubernetes.io/projected/189cb4b9-5c76-4872-b350-29525973b831-kube-api-access-jvfvv\") pod \"certified-operators-k4tjc\" (UID: \"189cb4b9-5c76-4872-b350-29525973b831\") " pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:50 crc kubenswrapper[4822]: I1124 15:14:50.069074 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/189cb4b9-5c76-4872-b350-29525973b831-utilities\") pod \"certified-operators-k4tjc\" (UID: \"189cb4b9-5c76-4872-b350-29525973b831\") " pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:50 crc kubenswrapper[4822]: I1124 15:14:50.069189 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/189cb4b9-5c76-4872-b350-29525973b831-catalog-content\") pod \"certified-operators-k4tjc\" (UID: \"189cb4b9-5c76-4872-b350-29525973b831\") " pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:50 crc kubenswrapper[4822]: I1124 15:14:50.069636 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/189cb4b9-5c76-4872-b350-29525973b831-utilities\") pod \"certified-operators-k4tjc\" (UID: \"189cb4b9-5c76-4872-b350-29525973b831\") " pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:50 crc kubenswrapper[4822]: I1124 15:14:50.069783 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/189cb4b9-5c76-4872-b350-29525973b831-catalog-content\") pod \"certified-operators-k4tjc\" (UID: \"189cb4b9-5c76-4872-b350-29525973b831\") " pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:50 crc kubenswrapper[4822]: I1124 15:14:50.111092 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvfvv\" (UniqueName: \"kubernetes.io/projected/189cb4b9-5c76-4872-b350-29525973b831-kube-api-access-jvfvv\") pod \"certified-operators-k4tjc\" (UID: \"189cb4b9-5c76-4872-b350-29525973b831\") " pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:50 crc kubenswrapper[4822]: I1124 15:14:50.213738 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:14:50 crc kubenswrapper[4822]: W1124 15:14:50.727611 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod189cb4b9_5c76_4872_b350_29525973b831.slice/crio-516ccc84ae626b6bcbcb86a326923e6d3ba0309e2ef84c0d6e1f9166758bb206 WatchSource:0}: Error finding container 516ccc84ae626b6bcbcb86a326923e6d3ba0309e2ef84c0d6e1f9166758bb206: Status 404 returned error can't find the container with id 516ccc84ae626b6bcbcb86a326923e6d3ba0309e2ef84c0d6e1f9166758bb206 Nov 24 15:14:50 crc kubenswrapper[4822]: I1124 15:14:50.732291 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k4tjc"] Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.097309 4822 generic.go:334] "Generic (PLEG): container finished" podID="189cb4b9-5c76-4872-b350-29525973b831" containerID="2209728c0b9d392f29bb83003f6457fb8e72f192d182db0215a3508b83898be0" exitCode=0 Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.097357 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4tjc" event={"ID":"189cb4b9-5c76-4872-b350-29525973b831","Type":"ContainerDied","Data":"2209728c0b9d392f29bb83003f6457fb8e72f192d182db0215a3508b83898be0"} Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.097557 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4tjc" event={"ID":"189cb4b9-5c76-4872-b350-29525973b831","Type":"ContainerStarted","Data":"516ccc84ae626b6bcbcb86a326923e6d3ba0309e2ef84c0d6e1f9166758bb206"} Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.099162 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.253148 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jwdch"] Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.255901 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.269662 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jwdch"] Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.397770 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-catalog-content\") pod \"community-operators-jwdch\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.397955 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfcwm\" (UniqueName: \"kubernetes.io/projected/f21c2f35-8023-4d87-b425-d682606e0028-kube-api-access-lfcwm\") pod \"community-operators-jwdch\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.398090 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-utilities\") pod \"community-operators-jwdch\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.500462 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-catalog-content\") pod \"community-operators-jwdch\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.500851 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfcwm\" (UniqueName: \"kubernetes.io/projected/f21c2f35-8023-4d87-b425-d682606e0028-kube-api-access-lfcwm\") pod \"community-operators-jwdch\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.501235 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-utilities\") pod \"community-operators-jwdch\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.501301 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-catalog-content\") pod \"community-operators-jwdch\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.501709 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-utilities\") pod \"community-operators-jwdch\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.526781 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfcwm\" (UniqueName: \"kubernetes.io/projected/f21c2f35-8023-4d87-b425-d682606e0028-kube-api-access-lfcwm\") pod \"community-operators-jwdch\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:51 crc kubenswrapper[4822]: I1124 15:14:51.582940 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:14:52 crc kubenswrapper[4822]: I1124 15:14:52.179484 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jwdch"] Nov 24 15:14:52 crc kubenswrapper[4822]: W1124 15:14:52.192191 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf21c2f35_8023_4d87_b425_d682606e0028.slice/crio-84ef40bf9f47c4386c734370b61135b4ca7774f5cc3cc16130c47f69d81b3867 WatchSource:0}: Error finding container 84ef40bf9f47c4386c734370b61135b4ca7774f5cc3cc16130c47f69d81b3867: Status 404 returned error can't find the container with id 84ef40bf9f47c4386c734370b61135b4ca7774f5cc3cc16130c47f69d81b3867 Nov 24 15:14:53 crc kubenswrapper[4822]: I1124 15:14:53.151817 4822 generic.go:334] "Generic (PLEG): container finished" podID="f21c2f35-8023-4d87-b425-d682606e0028" containerID="af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3" exitCode=0 Nov 24 15:14:53 crc kubenswrapper[4822]: I1124 15:14:53.152114 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwdch" event={"ID":"f21c2f35-8023-4d87-b425-d682606e0028","Type":"ContainerDied","Data":"af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3"} Nov 24 15:14:53 crc kubenswrapper[4822]: I1124 15:14:53.152140 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwdch" event={"ID":"f21c2f35-8023-4d87-b425-d682606e0028","Type":"ContainerStarted","Data":"84ef40bf9f47c4386c734370b61135b4ca7774f5cc3cc16130c47f69d81b3867"} Nov 24 15:14:54 crc kubenswrapper[4822]: I1124 15:14:54.163551 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwdch" event={"ID":"f21c2f35-8023-4d87-b425-d682606e0028","Type":"ContainerStarted","Data":"9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306"} Nov 24 15:14:55 crc kubenswrapper[4822]: I1124 15:14:55.176354 4822 generic.go:334] "Generic (PLEG): container finished" podID="f21c2f35-8023-4d87-b425-d682606e0028" containerID="9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306" exitCode=0 Nov 24 15:14:55 crc kubenswrapper[4822]: I1124 15:14:55.176448 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwdch" event={"ID":"f21c2f35-8023-4d87-b425-d682606e0028","Type":"ContainerDied","Data":"9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306"} Nov 24 15:14:57 crc kubenswrapper[4822]: I1124 15:14:57.203507 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4tjc" event={"ID":"189cb4b9-5c76-4872-b350-29525973b831","Type":"ContainerStarted","Data":"3eed4de72c9ddeefedb78ab43e40a5fc3dd5923785aa227a2738e90433468335"} Nov 24 15:14:58 crc kubenswrapper[4822]: I1124 15:14:58.222674 4822 generic.go:334] "Generic (PLEG): container finished" podID="189cb4b9-5c76-4872-b350-29525973b831" containerID="3eed4de72c9ddeefedb78ab43e40a5fc3dd5923785aa227a2738e90433468335" exitCode=0 Nov 24 15:14:58 crc kubenswrapper[4822]: I1124 15:14:58.222769 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4tjc" event={"ID":"189cb4b9-5c76-4872-b350-29525973b831","Type":"ContainerDied","Data":"3eed4de72c9ddeefedb78ab43e40a5fc3dd5923785aa227a2738e90433468335"} Nov 24 15:14:58 crc kubenswrapper[4822]: I1124 15:14:58.226074 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwdch" event={"ID":"f21c2f35-8023-4d87-b425-d682606e0028","Type":"ContainerStarted","Data":"1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43"} Nov 24 15:14:58 crc kubenswrapper[4822]: I1124 15:14:58.278571 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jwdch" podStartSLOduration=3.257484719 podStartE2EDuration="7.278546474s" podCreationTimestamp="2025-11-24 15:14:51 +0000 UTC" firstStartedPulling="2025-11-24 15:14:53.157026707 +0000 UTC m=+3330.273667184" lastFinishedPulling="2025-11-24 15:14:57.178088452 +0000 UTC m=+3334.294728939" observedRunningTime="2025-11-24 15:14:58.270743441 +0000 UTC m=+3335.387383988" watchObservedRunningTime="2025-11-24 15:14:58.278546474 +0000 UTC m=+3335.395186971" Nov 24 15:14:59 crc kubenswrapper[4822]: I1124 15:14:59.248726 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4tjc" event={"ID":"189cb4b9-5c76-4872-b350-29525973b831","Type":"ContainerStarted","Data":"59ea326afe45463f0eaad48eee99eb959e3f67b6fe1a3613aa4a6d8921b3b332"} Nov 24 15:14:59 crc kubenswrapper[4822]: I1124 15:14:59.281905 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k4tjc" podStartSLOduration=2.762021966 podStartE2EDuration="10.281885441s" podCreationTimestamp="2025-11-24 15:14:49 +0000 UTC" firstStartedPulling="2025-11-24 15:14:51.098941043 +0000 UTC m=+3328.215581520" lastFinishedPulling="2025-11-24 15:14:58.618804518 +0000 UTC m=+3335.735444995" observedRunningTime="2025-11-24 15:14:59.269712781 +0000 UTC m=+3336.386353258" watchObservedRunningTime="2025-11-24 15:14:59.281885441 +0000 UTC m=+3336.398525928" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.174720 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp"] Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.176348 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.178562 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.178688 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.188046 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp"] Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.214511 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.215841 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.334680 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smj72\" (UniqueName: \"kubernetes.io/projected/31d6a076-3cbd-4f33-9796-3f79768f3c1f-kube-api-access-smj72\") pod \"collect-profiles-29399955-57wvp\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.334928 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31d6a076-3cbd-4f33-9796-3f79768f3c1f-config-volume\") pod \"collect-profiles-29399955-57wvp\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.335100 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31d6a076-3cbd-4f33-9796-3f79768f3c1f-secret-volume\") pod \"collect-profiles-29399955-57wvp\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.437275 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smj72\" (UniqueName: \"kubernetes.io/projected/31d6a076-3cbd-4f33-9796-3f79768f3c1f-kube-api-access-smj72\") pod \"collect-profiles-29399955-57wvp\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.437384 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31d6a076-3cbd-4f33-9796-3f79768f3c1f-config-volume\") pod \"collect-profiles-29399955-57wvp\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.437413 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31d6a076-3cbd-4f33-9796-3f79768f3c1f-secret-volume\") pod \"collect-profiles-29399955-57wvp\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.438391 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31d6a076-3cbd-4f33-9796-3f79768f3c1f-config-volume\") pod \"collect-profiles-29399955-57wvp\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.442993 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31d6a076-3cbd-4f33-9796-3f79768f3c1f-secret-volume\") pod \"collect-profiles-29399955-57wvp\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.462460 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smj72\" (UniqueName: \"kubernetes.io/projected/31d6a076-3cbd-4f33-9796-3f79768f3c1f-kube-api-access-smj72\") pod \"collect-profiles-29399955-57wvp\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:00 crc kubenswrapper[4822]: I1124 15:15:00.516194 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:01 crc kubenswrapper[4822]: I1124 15:15:01.013638 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp"] Nov 24 15:15:01 crc kubenswrapper[4822]: I1124 15:15:01.274674 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-k4tjc" podUID="189cb4b9-5c76-4872-b350-29525973b831" containerName="registry-server" probeResult="failure" output=< Nov 24 15:15:01 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 15:15:01 crc kubenswrapper[4822]: > Nov 24 15:15:01 crc kubenswrapper[4822]: I1124 15:15:01.283709 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" event={"ID":"31d6a076-3cbd-4f33-9796-3f79768f3c1f","Type":"ContainerStarted","Data":"4b168a40e10689eed4682b926afa76377cb17968cfb3c75351317187174e3018"} Nov 24 15:15:01 crc kubenswrapper[4822]: I1124 15:15:01.283786 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" event={"ID":"31d6a076-3cbd-4f33-9796-3f79768f3c1f","Type":"ContainerStarted","Data":"bd876153b3d010f765ee1863f3fd40bab415b5ad1a285560a32f6b360411045c"} Nov 24 15:15:01 crc kubenswrapper[4822]: I1124 15:15:01.305405 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" podStartSLOduration=1.305385597 podStartE2EDuration="1.305385597s" podCreationTimestamp="2025-11-24 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:15:01.300457384 +0000 UTC m=+3338.417097871" watchObservedRunningTime="2025-11-24 15:15:01.305385597 +0000 UTC m=+3338.422026094" Nov 24 15:15:01 crc kubenswrapper[4822]: I1124 15:15:01.583854 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:15:01 crc kubenswrapper[4822]: I1124 15:15:01.583932 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:15:01 crc kubenswrapper[4822]: I1124 15:15:01.657979 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:15:02 crc kubenswrapper[4822]: I1124 15:15:02.307683 4822 generic.go:334] "Generic (PLEG): container finished" podID="31d6a076-3cbd-4f33-9796-3f79768f3c1f" containerID="4b168a40e10689eed4682b926afa76377cb17968cfb3c75351317187174e3018" exitCode=0 Nov 24 15:15:02 crc kubenswrapper[4822]: I1124 15:15:02.307812 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" event={"ID":"31d6a076-3cbd-4f33-9796-3f79768f3c1f","Type":"ContainerDied","Data":"4b168a40e10689eed4682b926afa76377cb17968cfb3c75351317187174e3018"} Nov 24 15:15:02 crc kubenswrapper[4822]: I1124 15:15:02.362426 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:15:02 crc kubenswrapper[4822]: I1124 15:15:02.652668 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jwdch"] Nov 24 15:15:03 crc kubenswrapper[4822]: I1124 15:15:03.764419 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:03 crc kubenswrapper[4822]: I1124 15:15:03.908463 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smj72\" (UniqueName: \"kubernetes.io/projected/31d6a076-3cbd-4f33-9796-3f79768f3c1f-kube-api-access-smj72\") pod \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " Nov 24 15:15:03 crc kubenswrapper[4822]: I1124 15:15:03.908759 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31d6a076-3cbd-4f33-9796-3f79768f3c1f-secret-volume\") pod \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " Nov 24 15:15:03 crc kubenswrapper[4822]: I1124 15:15:03.908804 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31d6a076-3cbd-4f33-9796-3f79768f3c1f-config-volume\") pod \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\" (UID: \"31d6a076-3cbd-4f33-9796-3f79768f3c1f\") " Nov 24 15:15:03 crc kubenswrapper[4822]: I1124 15:15:03.909601 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d6a076-3cbd-4f33-9796-3f79768f3c1f-config-volume" (OuterVolumeSpecName: "config-volume") pod "31d6a076-3cbd-4f33-9796-3f79768f3c1f" (UID: "31d6a076-3cbd-4f33-9796-3f79768f3c1f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:15:03 crc kubenswrapper[4822]: I1124 15:15:03.918419 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d6a076-3cbd-4f33-9796-3f79768f3c1f-kube-api-access-smj72" (OuterVolumeSpecName: "kube-api-access-smj72") pod "31d6a076-3cbd-4f33-9796-3f79768f3c1f" (UID: "31d6a076-3cbd-4f33-9796-3f79768f3c1f"). InnerVolumeSpecName "kube-api-access-smj72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:15:03 crc kubenswrapper[4822]: I1124 15:15:03.919961 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d6a076-3cbd-4f33-9796-3f79768f3c1f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "31d6a076-3cbd-4f33-9796-3f79768f3c1f" (UID: "31d6a076-3cbd-4f33-9796-3f79768f3c1f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.011916 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smj72\" (UniqueName: \"kubernetes.io/projected/31d6a076-3cbd-4f33-9796-3f79768f3c1f-kube-api-access-smj72\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.011947 4822 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31d6a076-3cbd-4f33-9796-3f79768f3c1f-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.011956 4822 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31d6a076-3cbd-4f33-9796-3f79768f3c1f-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.368525 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jwdch" podUID="f21c2f35-8023-4d87-b425-d682606e0028" containerName="registry-server" containerID="cri-o://1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43" gracePeriod=2 Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.368859 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" event={"ID":"31d6a076-3cbd-4f33-9796-3f79768f3c1f","Type":"ContainerDied","Data":"bd876153b3d010f765ee1863f3fd40bab415b5ad1a285560a32f6b360411045c"} Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.372030 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd876153b3d010f765ee1863f3fd40bab415b5ad1a285560a32f6b360411045c" Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.368939 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-57wvp" Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.411379 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv"] Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.418761 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-ljbfv"] Nov 24 15:15:04 crc kubenswrapper[4822]: I1124 15:15:04.915879 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.029420 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-catalog-content\") pod \"f21c2f35-8023-4d87-b425-d682606e0028\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.030062 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-utilities\") pod \"f21c2f35-8023-4d87-b425-d682606e0028\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.030564 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfcwm\" (UniqueName: \"kubernetes.io/projected/f21c2f35-8023-4d87-b425-d682606e0028-kube-api-access-lfcwm\") pod \"f21c2f35-8023-4d87-b425-d682606e0028\" (UID: \"f21c2f35-8023-4d87-b425-d682606e0028\") " Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.030865 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-utilities" (OuterVolumeSpecName: "utilities") pod "f21c2f35-8023-4d87-b425-d682606e0028" (UID: "f21c2f35-8023-4d87-b425-d682606e0028"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.031593 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.047654 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f21c2f35-8023-4d87-b425-d682606e0028-kube-api-access-lfcwm" (OuterVolumeSpecName: "kube-api-access-lfcwm") pod "f21c2f35-8023-4d87-b425-d682606e0028" (UID: "f21c2f35-8023-4d87-b425-d682606e0028"). InnerVolumeSpecName "kube-api-access-lfcwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.108664 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f21c2f35-8023-4d87-b425-d682606e0028" (UID: "f21c2f35-8023-4d87-b425-d682606e0028"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.133906 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfcwm\" (UniqueName: \"kubernetes.io/projected/f21c2f35-8023-4d87-b425-d682606e0028-kube-api-access-lfcwm\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.133961 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21c2f35-8023-4d87-b425-d682606e0028-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.384857 4822 generic.go:334] "Generic (PLEG): container finished" podID="f21c2f35-8023-4d87-b425-d682606e0028" containerID="1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43" exitCode=0 Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.384908 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwdch" event={"ID":"f21c2f35-8023-4d87-b425-d682606e0028","Type":"ContainerDied","Data":"1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43"} Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.384947 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwdch" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.384972 4822 scope.go:117] "RemoveContainer" containerID="1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.384954 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwdch" event={"ID":"f21c2f35-8023-4d87-b425-d682606e0028","Type":"ContainerDied","Data":"84ef40bf9f47c4386c734370b61135b4ca7774f5cc3cc16130c47f69d81b3867"} Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.417368 4822 scope.go:117] "RemoveContainer" containerID="9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.457708 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jwdch"] Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.461187 4822 scope.go:117] "RemoveContainer" containerID="af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.470924 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jwdch"] Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.511781 4822 scope.go:117] "RemoveContainer" containerID="1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43" Nov 24 15:15:05 crc kubenswrapper[4822]: E1124 15:15:05.512252 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43\": container with ID starting with 1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43 not found: ID does not exist" containerID="1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.512300 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43"} err="failed to get container status \"1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43\": rpc error: code = NotFound desc = could not find container \"1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43\": container with ID starting with 1f5d47e1f950420c3a919afbaf391c15b1a3af79f1e783126d6376ec0e065b43 not found: ID does not exist" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.512335 4822 scope.go:117] "RemoveContainer" containerID="9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306" Nov 24 15:15:05 crc kubenswrapper[4822]: E1124 15:15:05.512886 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306\": container with ID starting with 9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306 not found: ID does not exist" containerID="9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.512934 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306"} err="failed to get container status \"9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306\": rpc error: code = NotFound desc = could not find container \"9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306\": container with ID starting with 9f150708a73d85a0238103c76843df51fb1dbe0bcb0d01ba089f761148555306 not found: ID does not exist" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.512968 4822 scope.go:117] "RemoveContainer" containerID="af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3" Nov 24 15:15:05 crc kubenswrapper[4822]: E1124 15:15:05.513374 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3\": container with ID starting with af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3 not found: ID does not exist" containerID="af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.513419 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3"} err="failed to get container status \"af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3\": rpc error: code = NotFound desc = could not find container \"af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3\": container with ID starting with af640b70f0816d7581d720c01ea44850b4d646db90168973f60d268975c394d3 not found: ID does not exist" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.719916 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f21c2f35-8023-4d87-b425-d682606e0028" path="/var/lib/kubelet/pods/f21c2f35-8023-4d87-b425-d682606e0028/volumes" Nov 24 15:15:05 crc kubenswrapper[4822]: I1124 15:15:05.722661 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd909522-d5b8-4024-a3c5-f1d04821a578" path="/var/lib/kubelet/pods/fd909522-d5b8-4024-a3c5-f1d04821a578/volumes" Nov 24 15:15:10 crc kubenswrapper[4822]: I1124 15:15:10.282920 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:15:10 crc kubenswrapper[4822]: I1124 15:15:10.364963 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k4tjc" Nov 24 15:15:10 crc kubenswrapper[4822]: I1124 15:15:10.434468 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k4tjc"] Nov 24 15:15:10 crc kubenswrapper[4822]: I1124 15:15:10.525552 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-28rbg"] Nov 24 15:15:10 crc kubenswrapper[4822]: I1124 15:15:10.525790 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-28rbg" podUID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerName="registry-server" containerID="cri-o://d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0" gracePeriod=2 Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.083677 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28rbg" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.202986 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-catalog-content\") pod \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.203157 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wttpp\" (UniqueName: \"kubernetes.io/projected/5c76e331-f8e8-45e9-9afb-2add7327ebd1-kube-api-access-wttpp\") pod \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.203293 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-utilities\") pod \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\" (UID: \"5c76e331-f8e8-45e9-9afb-2add7327ebd1\") " Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.206361 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-utilities" (OuterVolumeSpecName: "utilities") pod "5c76e331-f8e8-45e9-9afb-2add7327ebd1" (UID: "5c76e331-f8e8-45e9-9afb-2add7327ebd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.209704 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c76e331-f8e8-45e9-9afb-2add7327ebd1-kube-api-access-wttpp" (OuterVolumeSpecName: "kube-api-access-wttpp") pod "5c76e331-f8e8-45e9-9afb-2add7327ebd1" (UID: "5c76e331-f8e8-45e9-9afb-2add7327ebd1"). InnerVolumeSpecName "kube-api-access-wttpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.281504 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c76e331-f8e8-45e9-9afb-2add7327ebd1" (UID: "5c76e331-f8e8-45e9-9afb-2add7327ebd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.306105 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.306225 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c76e331-f8e8-45e9-9afb-2add7327ebd1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.306249 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wttpp\" (UniqueName: \"kubernetes.io/projected/5c76e331-f8e8-45e9-9afb-2add7327ebd1-kube-api-access-wttpp\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.461984 4822 generic.go:334] "Generic (PLEG): container finished" podID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerID="d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0" exitCode=0 Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.462058 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28rbg" event={"ID":"5c76e331-f8e8-45e9-9afb-2add7327ebd1","Type":"ContainerDied","Data":"d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0"} Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.462088 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28rbg" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.462140 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28rbg" event={"ID":"5c76e331-f8e8-45e9-9afb-2add7327ebd1","Type":"ContainerDied","Data":"905e8c024cd3af654c57166b4e5014d38a1d1530af4233882cbbe8d329ebb5a8"} Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.462177 4822 scope.go:117] "RemoveContainer" containerID="d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.498775 4822 scope.go:117] "RemoveContainer" containerID="93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.519318 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-28rbg"] Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.535278 4822 scope.go:117] "RemoveContainer" containerID="7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.537291 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-28rbg"] Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.589634 4822 scope.go:117] "RemoveContainer" containerID="d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0" Nov 24 15:15:11 crc kubenswrapper[4822]: E1124 15:15:11.590049 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0\": container with ID starting with d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0 not found: ID does not exist" containerID="d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.590093 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0"} err="failed to get container status \"d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0\": rpc error: code = NotFound desc = could not find container \"d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0\": container with ID starting with d750aef4b856453c104717a76d5b4dc347d479738ce54be5510aa60d7cacbea0 not found: ID does not exist" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.590126 4822 scope.go:117] "RemoveContainer" containerID="93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0" Nov 24 15:15:11 crc kubenswrapper[4822]: E1124 15:15:11.590788 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0\": container with ID starting with 93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0 not found: ID does not exist" containerID="93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.590820 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0"} err="failed to get container status \"93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0\": rpc error: code = NotFound desc = could not find container \"93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0\": container with ID starting with 93362e81b80f4d03f4b74fccd62326e90e3e051f48ce41d4bc4101cc25d5b1f0 not found: ID does not exist" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.590838 4822 scope.go:117] "RemoveContainer" containerID="7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad" Nov 24 15:15:11 crc kubenswrapper[4822]: E1124 15:15:11.592443 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad\": container with ID starting with 7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad not found: ID does not exist" containerID="7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.592487 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad"} err="failed to get container status \"7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad\": rpc error: code = NotFound desc = could not find container \"7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad\": container with ID starting with 7370e83c8d1ea2e2e28f74bde8ebcbe89557fc46a41a90d583c2a46ddbe67bad not found: ID does not exist" Nov 24 15:15:11 crc kubenswrapper[4822]: I1124 15:15:11.742932 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" path="/var/lib/kubelet/pods/5c76e331-f8e8-45e9-9afb-2add7327ebd1/volumes" Nov 24 15:15:30 crc kubenswrapper[4822]: I1124 15:15:30.047854 4822 scope.go:117] "RemoveContainer" containerID="c9de05de335c3cd8837277fbae4a5826a2f3c0721dd9e5338f7dfaaaa3e5ea71" Nov 24 15:15:35 crc kubenswrapper[4822]: I1124 15:15:35.756744 4822 generic.go:334] "Generic (PLEG): container finished" podID="561e27e7-d8ea-4764-872e-0b022659275c" containerID="34d88b00b7f3fe35002b6add35967d052817b15de64c99d0cf68ae185094ae90" exitCode=0 Nov 24 15:15:35 crc kubenswrapper[4822]: I1124 15:15:35.756855 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" event={"ID":"561e27e7-d8ea-4764-872e-0b022659275c","Type":"ContainerDied","Data":"34d88b00b7f3fe35002b6add35967d052817b15de64c99d0cf68ae185094ae90"} Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.350749 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.435005 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ssh-key\") pod \"561e27e7-d8ea-4764-872e-0b022659275c\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.435107 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bthgd\" (UniqueName: \"kubernetes.io/projected/561e27e7-d8ea-4764-872e-0b022659275c-kube-api-access-bthgd\") pod \"561e27e7-d8ea-4764-872e-0b022659275c\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.435320 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-inventory\") pod \"561e27e7-d8ea-4764-872e-0b022659275c\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.435485 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-2\") pod \"561e27e7-d8ea-4764-872e-0b022659275c\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.435618 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-telemetry-combined-ca-bundle\") pod \"561e27e7-d8ea-4764-872e-0b022659275c\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.435756 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-0\") pod \"561e27e7-d8ea-4764-872e-0b022659275c\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.435812 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-1\") pod \"561e27e7-d8ea-4764-872e-0b022659275c\" (UID: \"561e27e7-d8ea-4764-872e-0b022659275c\") " Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.441065 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "561e27e7-d8ea-4764-872e-0b022659275c" (UID: "561e27e7-d8ea-4764-872e-0b022659275c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.444169 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561e27e7-d8ea-4764-872e-0b022659275c-kube-api-access-bthgd" (OuterVolumeSpecName: "kube-api-access-bthgd") pod "561e27e7-d8ea-4764-872e-0b022659275c" (UID: "561e27e7-d8ea-4764-872e-0b022659275c"). InnerVolumeSpecName "kube-api-access-bthgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.465222 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-inventory" (OuterVolumeSpecName: "inventory") pod "561e27e7-d8ea-4764-872e-0b022659275c" (UID: "561e27e7-d8ea-4764-872e-0b022659275c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.470033 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "561e27e7-d8ea-4764-872e-0b022659275c" (UID: "561e27e7-d8ea-4764-872e-0b022659275c"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.473105 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "561e27e7-d8ea-4764-872e-0b022659275c" (UID: "561e27e7-d8ea-4764-872e-0b022659275c"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.497125 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "561e27e7-d8ea-4764-872e-0b022659275c" (UID: "561e27e7-d8ea-4764-872e-0b022659275c"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.503849 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "561e27e7-d8ea-4764-872e-0b022659275c" (UID: "561e27e7-d8ea-4764-872e-0b022659275c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.542388 4822 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.542444 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bthgd\" (UniqueName: \"kubernetes.io/projected/561e27e7-d8ea-4764-872e-0b022659275c-kube-api-access-bthgd\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.542477 4822 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.542501 4822 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.542519 4822 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.542545 4822 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.542564 4822 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/561e27e7-d8ea-4764-872e-0b022659275c-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.784832 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" event={"ID":"561e27e7-d8ea-4764-872e-0b022659275c","Type":"ContainerDied","Data":"a0a1d2bc7c3f588486fa3b49034234f2700f9e8cb1ded4887568607420a03886"} Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.784879 4822 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0a1d2bc7c3f588486fa3b49034234f2700f9e8cb1ded4887568607420a03886" Nov 24 15:15:37 crc kubenswrapper[4822]: I1124 15:15:37.784969 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s22zg" Nov 24 15:15:41 crc kubenswrapper[4822]: I1124 15:15:41.577947 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:15:41 crc kubenswrapper[4822]: I1124 15:15:41.578434 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:16:11 crc kubenswrapper[4822]: I1124 15:16:11.578361 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:16:11 crc kubenswrapper[4822]: I1124 15:16:11.578975 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:16:41 crc kubenswrapper[4822]: I1124 15:16:41.578079 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:16:41 crc kubenswrapper[4822]: I1124 15:16:41.579092 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:16:41 crc kubenswrapper[4822]: I1124 15:16:41.579188 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 15:16:41 crc kubenswrapper[4822]: I1124 15:16:41.580491 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b89a251fd0f0cfe8d88737b003ae9166ecdc96e5233f626b16b042314ed274c"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:16:41 crc kubenswrapper[4822]: I1124 15:16:41.580617 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://7b89a251fd0f0cfe8d88737b003ae9166ecdc96e5233f626b16b042314ed274c" gracePeriod=600 Nov 24 15:16:42 crc kubenswrapper[4822]: I1124 15:16:42.640484 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="7b89a251fd0f0cfe8d88737b003ae9166ecdc96e5233f626b16b042314ed274c" exitCode=0 Nov 24 15:16:42 crc kubenswrapper[4822]: I1124 15:16:42.640521 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"7b89a251fd0f0cfe8d88737b003ae9166ecdc96e5233f626b16b042314ed274c"} Nov 24 15:16:42 crc kubenswrapper[4822]: I1124 15:16:42.641339 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5"} Nov 24 15:16:42 crc kubenswrapper[4822]: I1124 15:16:42.641382 4822 scope.go:117] "RemoveContainer" containerID="669ccf15f8587464461c934a30cdc7c3ee750b6e0586160bc853fef1db871102" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.965278 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-94kzw/must-gather-l9b7n"] Nov 24 15:17:04 crc kubenswrapper[4822]: E1124 15:17:04.966234 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerName="registry-server" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966249 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerName="registry-server" Nov 24 15:17:04 crc kubenswrapper[4822]: E1124 15:17:04.966268 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f21c2f35-8023-4d87-b425-d682606e0028" containerName="registry-server" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966274 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f21c2f35-8023-4d87-b425-d682606e0028" containerName="registry-server" Nov 24 15:17:04 crc kubenswrapper[4822]: E1124 15:17:04.966284 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d6a076-3cbd-4f33-9796-3f79768f3c1f" containerName="collect-profiles" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966290 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d6a076-3cbd-4f33-9796-3f79768f3c1f" containerName="collect-profiles" Nov 24 15:17:04 crc kubenswrapper[4822]: E1124 15:17:04.966310 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerName="extract-utilities" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966316 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerName="extract-utilities" Nov 24 15:17:04 crc kubenswrapper[4822]: E1124 15:17:04.966330 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f21c2f35-8023-4d87-b425-d682606e0028" containerName="extract-content" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966335 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f21c2f35-8023-4d87-b425-d682606e0028" containerName="extract-content" Nov 24 15:17:04 crc kubenswrapper[4822]: E1124 15:17:04.966354 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561e27e7-d8ea-4764-872e-0b022659275c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966361 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="561e27e7-d8ea-4764-872e-0b022659275c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 15:17:04 crc kubenswrapper[4822]: E1124 15:17:04.966382 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f21c2f35-8023-4d87-b425-d682606e0028" containerName="extract-utilities" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966387 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="f21c2f35-8023-4d87-b425-d682606e0028" containerName="extract-utilities" Nov 24 15:17:04 crc kubenswrapper[4822]: E1124 15:17:04.966401 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerName="extract-content" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966407 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerName="extract-content" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966618 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="f21c2f35-8023-4d87-b425-d682606e0028" containerName="registry-server" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966641 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="561e27e7-d8ea-4764-872e-0b022659275c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966654 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d6a076-3cbd-4f33-9796-3f79768f3c1f" containerName="collect-profiles" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.966670 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c76e331-f8e8-45e9-9afb-2add7327ebd1" containerName="registry-server" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.968106 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.970955 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-94kzw"/"kube-root-ca.crt" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.971074 4822 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-94kzw"/"default-dockercfg-jk92k" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.971309 4822 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-94kzw"/"openshift-service-ca.crt" Nov 24 15:17:04 crc kubenswrapper[4822]: I1124 15:17:04.977432 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-94kzw/must-gather-l9b7n"] Nov 24 15:17:05 crc kubenswrapper[4822]: I1124 15:17:05.161588 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b810e4f5-1dc3-4713-8a8c-18d816697cef-must-gather-output\") pod \"must-gather-l9b7n\" (UID: \"b810e4f5-1dc3-4713-8a8c-18d816697cef\") " pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:17:05 crc kubenswrapper[4822]: I1124 15:17:05.161890 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7wms\" (UniqueName: \"kubernetes.io/projected/b810e4f5-1dc3-4713-8a8c-18d816697cef-kube-api-access-n7wms\") pod \"must-gather-l9b7n\" (UID: \"b810e4f5-1dc3-4713-8a8c-18d816697cef\") " pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:17:05 crc kubenswrapper[4822]: I1124 15:17:05.264019 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b810e4f5-1dc3-4713-8a8c-18d816697cef-must-gather-output\") pod \"must-gather-l9b7n\" (UID: \"b810e4f5-1dc3-4713-8a8c-18d816697cef\") " pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:17:05 crc kubenswrapper[4822]: I1124 15:17:05.264067 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7wms\" (UniqueName: \"kubernetes.io/projected/b810e4f5-1dc3-4713-8a8c-18d816697cef-kube-api-access-n7wms\") pod \"must-gather-l9b7n\" (UID: \"b810e4f5-1dc3-4713-8a8c-18d816697cef\") " pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:17:05 crc kubenswrapper[4822]: I1124 15:17:05.264594 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b810e4f5-1dc3-4713-8a8c-18d816697cef-must-gather-output\") pod \"must-gather-l9b7n\" (UID: \"b810e4f5-1dc3-4713-8a8c-18d816697cef\") " pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:17:05 crc kubenswrapper[4822]: I1124 15:17:05.290244 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7wms\" (UniqueName: \"kubernetes.io/projected/b810e4f5-1dc3-4713-8a8c-18d816697cef-kube-api-access-n7wms\") pod \"must-gather-l9b7n\" (UID: \"b810e4f5-1dc3-4713-8a8c-18d816697cef\") " pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:17:05 crc kubenswrapper[4822]: I1124 15:17:05.290965 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:17:05 crc kubenswrapper[4822]: I1124 15:17:05.766038 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-94kzw/must-gather-l9b7n"] Nov 24 15:17:05 crc kubenswrapper[4822]: W1124 15:17:05.777329 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb810e4f5_1dc3_4713_8a8c_18d816697cef.slice/crio-14cadaa43d7125964d55ddac36be5a030eb01f84ac86e96d8ad6b9bbdea632fa WatchSource:0}: Error finding container 14cadaa43d7125964d55ddac36be5a030eb01f84ac86e96d8ad6b9bbdea632fa: Status 404 returned error can't find the container with id 14cadaa43d7125964d55ddac36be5a030eb01f84ac86e96d8ad6b9bbdea632fa Nov 24 15:17:05 crc kubenswrapper[4822]: I1124 15:17:05.959387 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/must-gather-l9b7n" event={"ID":"b810e4f5-1dc3-4713-8a8c-18d816697cef","Type":"ContainerStarted","Data":"14cadaa43d7125964d55ddac36be5a030eb01f84ac86e96d8ad6b9bbdea632fa"} Nov 24 15:17:11 crc kubenswrapper[4822]: I1124 15:17:11.012067 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/must-gather-l9b7n" event={"ID":"b810e4f5-1dc3-4713-8a8c-18d816697cef","Type":"ContainerStarted","Data":"0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6"} Nov 24 15:17:11 crc kubenswrapper[4822]: I1124 15:17:11.012601 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/must-gather-l9b7n" event={"ID":"b810e4f5-1dc3-4713-8a8c-18d816697cef","Type":"ContainerStarted","Data":"9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c"} Nov 24 15:17:11 crc kubenswrapper[4822]: I1124 15:17:11.030599 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-94kzw/must-gather-l9b7n" podStartSLOduration=3.010631183 podStartE2EDuration="7.030569293s" podCreationTimestamp="2025-11-24 15:17:04 +0000 UTC" firstStartedPulling="2025-11-24 15:17:05.778729034 +0000 UTC m=+3462.895369501" lastFinishedPulling="2025-11-24 15:17:09.798667134 +0000 UTC m=+3466.915307611" observedRunningTime="2025-11-24 15:17:11.024333198 +0000 UTC m=+3468.140973685" watchObservedRunningTime="2025-11-24 15:17:11.030569293 +0000 UTC m=+3468.147209780" Nov 24 15:17:12 crc kubenswrapper[4822]: E1124 15:17:12.286457 4822 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.243:40676->38.102.83.243:33565: write tcp 38.102.83.243:40676->38.102.83.243:33565: write: broken pipe Nov 24 15:17:13 crc kubenswrapper[4822]: I1124 15:17:13.973295 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-94kzw/crc-debug-qzbfm"] Nov 24 15:17:13 crc kubenswrapper[4822]: I1124 15:17:13.975116 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:14 crc kubenswrapper[4822]: I1124 15:17:14.053719 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htd2v\" (UniqueName: \"kubernetes.io/projected/5f15cb73-df86-4d20-9fea-e7aa86d56f58-kube-api-access-htd2v\") pod \"crc-debug-qzbfm\" (UID: \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\") " pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:14 crc kubenswrapper[4822]: I1124 15:17:14.053815 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f15cb73-df86-4d20-9fea-e7aa86d56f58-host\") pod \"crc-debug-qzbfm\" (UID: \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\") " pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:14 crc kubenswrapper[4822]: I1124 15:17:14.155961 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htd2v\" (UniqueName: \"kubernetes.io/projected/5f15cb73-df86-4d20-9fea-e7aa86d56f58-kube-api-access-htd2v\") pod \"crc-debug-qzbfm\" (UID: \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\") " pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:14 crc kubenswrapper[4822]: I1124 15:17:14.156614 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f15cb73-df86-4d20-9fea-e7aa86d56f58-host\") pod \"crc-debug-qzbfm\" (UID: \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\") " pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:14 crc kubenswrapper[4822]: I1124 15:17:14.156693 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f15cb73-df86-4d20-9fea-e7aa86d56f58-host\") pod \"crc-debug-qzbfm\" (UID: \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\") " pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:14 crc kubenswrapper[4822]: I1124 15:17:14.173319 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htd2v\" (UniqueName: \"kubernetes.io/projected/5f15cb73-df86-4d20-9fea-e7aa86d56f58-kube-api-access-htd2v\") pod \"crc-debug-qzbfm\" (UID: \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\") " pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:14 crc kubenswrapper[4822]: I1124 15:17:14.302681 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:14 crc kubenswrapper[4822]: W1124 15:17:14.330033 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f15cb73_df86_4d20_9fea_e7aa86d56f58.slice/crio-4cf38f069fc018c0cb5c0ede7692fbd8a16410904f6239d9667dcfdb1e150112 WatchSource:0}: Error finding container 4cf38f069fc018c0cb5c0ede7692fbd8a16410904f6239d9667dcfdb1e150112: Status 404 returned error can't find the container with id 4cf38f069fc018c0cb5c0ede7692fbd8a16410904f6239d9667dcfdb1e150112 Nov 24 15:17:15 crc kubenswrapper[4822]: I1124 15:17:15.053270 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/crc-debug-qzbfm" event={"ID":"5f15cb73-df86-4d20-9fea-e7aa86d56f58","Type":"ContainerStarted","Data":"4cf38f069fc018c0cb5c0ede7692fbd8a16410904f6239d9667dcfdb1e150112"} Nov 24 15:17:25 crc kubenswrapper[4822]: I1124 15:17:25.161734 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/crc-debug-qzbfm" event={"ID":"5f15cb73-df86-4d20-9fea-e7aa86d56f58","Type":"ContainerStarted","Data":"51e24eb0fbd33724c60da974584d45dac7bd585e317ab73b1cac098c484eb847"} Nov 24 15:17:25 crc kubenswrapper[4822]: I1124 15:17:25.176587 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-94kzw/crc-debug-qzbfm" podStartSLOduration=1.588238317 podStartE2EDuration="12.176569103s" podCreationTimestamp="2025-11-24 15:17:13 +0000 UTC" firstStartedPulling="2025-11-24 15:17:14.332116545 +0000 UTC m=+3471.448757012" lastFinishedPulling="2025-11-24 15:17:24.920447321 +0000 UTC m=+3482.037087798" observedRunningTime="2025-11-24 15:17:25.174960882 +0000 UTC m=+3482.291601369" watchObservedRunningTime="2025-11-24 15:17:25.176569103 +0000 UTC m=+3482.293209580" Nov 24 15:17:44 crc kubenswrapper[4822]: I1124 15:17:44.358247 4822 generic.go:334] "Generic (PLEG): container finished" podID="5f15cb73-df86-4d20-9fea-e7aa86d56f58" containerID="51e24eb0fbd33724c60da974584d45dac7bd585e317ab73b1cac098c484eb847" exitCode=0 Nov 24 15:17:44 crc kubenswrapper[4822]: I1124 15:17:44.358734 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/crc-debug-qzbfm" event={"ID":"5f15cb73-df86-4d20-9fea-e7aa86d56f58","Type":"ContainerDied","Data":"51e24eb0fbd33724c60da974584d45dac7bd585e317ab73b1cac098c484eb847"} Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.505752 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.550999 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-94kzw/crc-debug-qzbfm"] Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.565597 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-94kzw/crc-debug-qzbfm"] Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.673550 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f15cb73-df86-4d20-9fea-e7aa86d56f58-host\") pod \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\" (UID: \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\") " Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.673672 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f15cb73-df86-4d20-9fea-e7aa86d56f58-host" (OuterVolumeSpecName: "host") pod "5f15cb73-df86-4d20-9fea-e7aa86d56f58" (UID: "5f15cb73-df86-4d20-9fea-e7aa86d56f58"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.673752 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htd2v\" (UniqueName: \"kubernetes.io/projected/5f15cb73-df86-4d20-9fea-e7aa86d56f58-kube-api-access-htd2v\") pod \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\" (UID: \"5f15cb73-df86-4d20-9fea-e7aa86d56f58\") " Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.674433 4822 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f15cb73-df86-4d20-9fea-e7aa86d56f58-host\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.683452 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f15cb73-df86-4d20-9fea-e7aa86d56f58-kube-api-access-htd2v" (OuterVolumeSpecName: "kube-api-access-htd2v") pod "5f15cb73-df86-4d20-9fea-e7aa86d56f58" (UID: "5f15cb73-df86-4d20-9fea-e7aa86d56f58"). InnerVolumeSpecName "kube-api-access-htd2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.716074 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f15cb73-df86-4d20-9fea-e7aa86d56f58" path="/var/lib/kubelet/pods/5f15cb73-df86-4d20-9fea-e7aa86d56f58/volumes" Nov 24 15:17:45 crc kubenswrapper[4822]: I1124 15:17:45.776129 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htd2v\" (UniqueName: \"kubernetes.io/projected/5f15cb73-df86-4d20-9fea-e7aa86d56f58-kube-api-access-htd2v\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.381498 4822 scope.go:117] "RemoveContainer" containerID="51e24eb0fbd33724c60da974584d45dac7bd585e317ab73b1cac098c484eb847" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.381668 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/crc-debug-qzbfm" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.765524 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-94kzw/crc-debug-5jrvb"] Nov 24 15:17:46 crc kubenswrapper[4822]: E1124 15:17:46.766130 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f15cb73-df86-4d20-9fea-e7aa86d56f58" containerName="container-00" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.766140 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f15cb73-df86-4d20-9fea-e7aa86d56f58" containerName="container-00" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.766341 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f15cb73-df86-4d20-9fea-e7aa86d56f58" containerName="container-00" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.767021 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.897044 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c84da9d2-6eeb-4f12-800e-b7aae9db848c-host\") pod \"crc-debug-5jrvb\" (UID: \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\") " pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.897121 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlvft\" (UniqueName: \"kubernetes.io/projected/c84da9d2-6eeb-4f12-800e-b7aae9db848c-kube-api-access-nlvft\") pod \"crc-debug-5jrvb\" (UID: \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\") " pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.999472 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c84da9d2-6eeb-4f12-800e-b7aae9db848c-host\") pod \"crc-debug-5jrvb\" (UID: \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\") " pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.999536 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlvft\" (UniqueName: \"kubernetes.io/projected/c84da9d2-6eeb-4f12-800e-b7aae9db848c-kube-api-access-nlvft\") pod \"crc-debug-5jrvb\" (UID: \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\") " pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:46 crc kubenswrapper[4822]: I1124 15:17:46.999656 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c84da9d2-6eeb-4f12-800e-b7aae9db848c-host\") pod \"crc-debug-5jrvb\" (UID: \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\") " pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:47 crc kubenswrapper[4822]: I1124 15:17:47.024030 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlvft\" (UniqueName: \"kubernetes.io/projected/c84da9d2-6eeb-4f12-800e-b7aae9db848c-kube-api-access-nlvft\") pod \"crc-debug-5jrvb\" (UID: \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\") " pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:47 crc kubenswrapper[4822]: I1124 15:17:47.082544 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:47 crc kubenswrapper[4822]: W1124 15:17:47.131973 4822 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc84da9d2_6eeb_4f12_800e_b7aae9db848c.slice/crio-18c78d1faead8cc10ba79d272428bf4b88687d9a512c480a280eafead5d332d7 WatchSource:0}: Error finding container 18c78d1faead8cc10ba79d272428bf4b88687d9a512c480a280eafead5d332d7: Status 404 returned error can't find the container with id 18c78d1faead8cc10ba79d272428bf4b88687d9a512c480a280eafead5d332d7 Nov 24 15:17:47 crc kubenswrapper[4822]: I1124 15:17:47.396082 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/crc-debug-5jrvb" event={"ID":"c84da9d2-6eeb-4f12-800e-b7aae9db848c","Type":"ContainerStarted","Data":"4684560e36c9690b325067821f1f3f1aee7088ed6c219f17764d5f0ce36920d9"} Nov 24 15:17:47 crc kubenswrapper[4822]: I1124 15:17:47.396488 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/crc-debug-5jrvb" event={"ID":"c84da9d2-6eeb-4f12-800e-b7aae9db848c","Type":"ContainerStarted","Data":"18c78d1faead8cc10ba79d272428bf4b88687d9a512c480a280eafead5d332d7"} Nov 24 15:17:47 crc kubenswrapper[4822]: I1124 15:17:47.410817 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-94kzw/crc-debug-5jrvb" podStartSLOduration=1.41080305 podStartE2EDuration="1.41080305s" podCreationTimestamp="2025-11-24 15:17:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:17:47.407444775 +0000 UTC m=+3504.524085272" watchObservedRunningTime="2025-11-24 15:17:47.41080305 +0000 UTC m=+3504.527443527" Nov 24 15:17:48 crc kubenswrapper[4822]: I1124 15:17:48.410925 4822 generic.go:334] "Generic (PLEG): container finished" podID="c84da9d2-6eeb-4f12-800e-b7aae9db848c" containerID="4684560e36c9690b325067821f1f3f1aee7088ed6c219f17764d5f0ce36920d9" exitCode=1 Nov 24 15:17:48 crc kubenswrapper[4822]: I1124 15:17:48.410963 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/crc-debug-5jrvb" event={"ID":"c84da9d2-6eeb-4f12-800e-b7aae9db848c","Type":"ContainerDied","Data":"4684560e36c9690b325067821f1f3f1aee7088ed6c219f17764d5f0ce36920d9"} Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.560869 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.593181 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-94kzw/crc-debug-5jrvb"] Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.602937 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-94kzw/crc-debug-5jrvb"] Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.656278 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c84da9d2-6eeb-4f12-800e-b7aae9db848c-host\") pod \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\" (UID: \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\") " Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.656416 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlvft\" (UniqueName: \"kubernetes.io/projected/c84da9d2-6eeb-4f12-800e-b7aae9db848c-kube-api-access-nlvft\") pod \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\" (UID: \"c84da9d2-6eeb-4f12-800e-b7aae9db848c\") " Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.656425 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c84da9d2-6eeb-4f12-800e-b7aae9db848c-host" (OuterVolumeSpecName: "host") pod "c84da9d2-6eeb-4f12-800e-b7aae9db848c" (UID: "c84da9d2-6eeb-4f12-800e-b7aae9db848c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.656881 4822 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c84da9d2-6eeb-4f12-800e-b7aae9db848c-host\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.661263 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c84da9d2-6eeb-4f12-800e-b7aae9db848c-kube-api-access-nlvft" (OuterVolumeSpecName: "kube-api-access-nlvft") pod "c84da9d2-6eeb-4f12-800e-b7aae9db848c" (UID: "c84da9d2-6eeb-4f12-800e-b7aae9db848c"). InnerVolumeSpecName "kube-api-access-nlvft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.714618 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c84da9d2-6eeb-4f12-800e-b7aae9db848c" path="/var/lib/kubelet/pods/c84da9d2-6eeb-4f12-800e-b7aae9db848c/volumes" Nov 24 15:17:49 crc kubenswrapper[4822]: I1124 15:17:49.759272 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlvft\" (UniqueName: \"kubernetes.io/projected/c84da9d2-6eeb-4f12-800e-b7aae9db848c-kube-api-access-nlvft\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:50 crc kubenswrapper[4822]: I1124 15:17:50.439536 4822 scope.go:117] "RemoveContainer" containerID="4684560e36c9690b325067821f1f3f1aee7088ed6c219f17764d5f0ce36920d9" Nov 24 15:17:50 crc kubenswrapper[4822]: I1124 15:17:50.439571 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/crc-debug-5jrvb" Nov 24 15:17:51 crc kubenswrapper[4822]: I1124 15:17:51.811982 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pqz8p"] Nov 24 15:17:51 crc kubenswrapper[4822]: E1124 15:17:51.812595 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c84da9d2-6eeb-4f12-800e-b7aae9db848c" containerName="container-00" Nov 24 15:17:51 crc kubenswrapper[4822]: I1124 15:17:51.812609 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="c84da9d2-6eeb-4f12-800e-b7aae9db848c" containerName="container-00" Nov 24 15:17:51 crc kubenswrapper[4822]: I1124 15:17:51.812820 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="c84da9d2-6eeb-4f12-800e-b7aae9db848c" containerName="container-00" Nov 24 15:17:51 crc kubenswrapper[4822]: I1124 15:17:51.814271 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:51 crc kubenswrapper[4822]: I1124 15:17:51.824910 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqz8p"] Nov 24 15:17:51 crc kubenswrapper[4822]: I1124 15:17:51.898545 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk2kd\" (UniqueName: \"kubernetes.io/projected/3d4268b7-60e8-42ff-b0d2-dac7a000170e-kube-api-access-dk2kd\") pod \"redhat-operators-pqz8p\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:51 crc kubenswrapper[4822]: I1124 15:17:51.898595 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-utilities\") pod \"redhat-operators-pqz8p\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:51 crc kubenswrapper[4822]: I1124 15:17:51.898618 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-catalog-content\") pod \"redhat-operators-pqz8p\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:52 crc kubenswrapper[4822]: I1124 15:17:52.000268 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-utilities\") pod \"redhat-operators-pqz8p\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:52 crc kubenswrapper[4822]: I1124 15:17:52.000329 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-catalog-content\") pod \"redhat-operators-pqz8p\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:52 crc kubenswrapper[4822]: I1124 15:17:52.000512 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk2kd\" (UniqueName: \"kubernetes.io/projected/3d4268b7-60e8-42ff-b0d2-dac7a000170e-kube-api-access-dk2kd\") pod \"redhat-operators-pqz8p\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:52 crc kubenswrapper[4822]: I1124 15:17:52.002113 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-utilities\") pod \"redhat-operators-pqz8p\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:52 crc kubenswrapper[4822]: I1124 15:17:52.002157 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-catalog-content\") pod \"redhat-operators-pqz8p\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:52 crc kubenswrapper[4822]: I1124 15:17:52.024384 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk2kd\" (UniqueName: \"kubernetes.io/projected/3d4268b7-60e8-42ff-b0d2-dac7a000170e-kube-api-access-dk2kd\") pod \"redhat-operators-pqz8p\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:52 crc kubenswrapper[4822]: I1124 15:17:52.138032 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:17:52 crc kubenswrapper[4822]: I1124 15:17:52.659072 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqz8p"] Nov 24 15:17:53 crc kubenswrapper[4822]: I1124 15:17:53.474548 4822 generic.go:334] "Generic (PLEG): container finished" podID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerID="050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8" exitCode=0 Nov 24 15:17:53 crc kubenswrapper[4822]: I1124 15:17:53.474642 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqz8p" event={"ID":"3d4268b7-60e8-42ff-b0d2-dac7a000170e","Type":"ContainerDied","Data":"050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8"} Nov 24 15:17:53 crc kubenswrapper[4822]: I1124 15:17:53.474847 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqz8p" event={"ID":"3d4268b7-60e8-42ff-b0d2-dac7a000170e","Type":"ContainerStarted","Data":"0fc166611b41a78adb56b719cefa4c45c93165ce1560f1767fb8379189812c25"} Nov 24 15:17:55 crc kubenswrapper[4822]: I1124 15:17:55.498896 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqz8p" event={"ID":"3d4268b7-60e8-42ff-b0d2-dac7a000170e","Type":"ContainerStarted","Data":"f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd"} Nov 24 15:17:58 crc kubenswrapper[4822]: I1124 15:17:58.555729 4822 generic.go:334] "Generic (PLEG): container finished" podID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerID="f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd" exitCode=0 Nov 24 15:17:58 crc kubenswrapper[4822]: I1124 15:17:58.555830 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqz8p" event={"ID":"3d4268b7-60e8-42ff-b0d2-dac7a000170e","Type":"ContainerDied","Data":"f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd"} Nov 24 15:17:59 crc kubenswrapper[4822]: I1124 15:17:59.574346 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqz8p" event={"ID":"3d4268b7-60e8-42ff-b0d2-dac7a000170e","Type":"ContainerStarted","Data":"ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505"} Nov 24 15:17:59 crc kubenswrapper[4822]: I1124 15:17:59.604205 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pqz8p" podStartSLOduration=3.071353116 podStartE2EDuration="8.604179062s" podCreationTimestamp="2025-11-24 15:17:51 +0000 UTC" firstStartedPulling="2025-11-24 15:17:53.476472729 +0000 UTC m=+3510.593113206" lastFinishedPulling="2025-11-24 15:17:59.009298675 +0000 UTC m=+3516.125939152" observedRunningTime="2025-11-24 15:17:59.592396306 +0000 UTC m=+3516.709036803" watchObservedRunningTime="2025-11-24 15:17:59.604179062 +0000 UTC m=+3516.720819569" Nov 24 15:18:02 crc kubenswrapper[4822]: I1124 15:18:02.138860 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:18:02 crc kubenswrapper[4822]: I1124 15:18:02.139226 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:18:03 crc kubenswrapper[4822]: I1124 15:18:03.222605 4822 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pqz8p" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerName="registry-server" probeResult="failure" output=< Nov 24 15:18:03 crc kubenswrapper[4822]: timeout: failed to connect service ":50051" within 1s Nov 24 15:18:03 crc kubenswrapper[4822]: > Nov 24 15:18:12 crc kubenswrapper[4822]: I1124 15:18:12.197085 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:18:12 crc kubenswrapper[4822]: I1124 15:18:12.261159 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:18:12 crc kubenswrapper[4822]: I1124 15:18:12.433894 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pqz8p"] Nov 24 15:18:13 crc kubenswrapper[4822]: I1124 15:18:13.739756 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pqz8p" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerName="registry-server" containerID="cri-o://ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505" gracePeriod=2 Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.323474 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.400968 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-utilities\") pod \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.401040 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk2kd\" (UniqueName: \"kubernetes.io/projected/3d4268b7-60e8-42ff-b0d2-dac7a000170e-kube-api-access-dk2kd\") pod \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.401324 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-catalog-content\") pod \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\" (UID: \"3d4268b7-60e8-42ff-b0d2-dac7a000170e\") " Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.402306 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-utilities" (OuterVolumeSpecName: "utilities") pod "3d4268b7-60e8-42ff-b0d2-dac7a000170e" (UID: "3d4268b7-60e8-42ff-b0d2-dac7a000170e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.414522 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d4268b7-60e8-42ff-b0d2-dac7a000170e-kube-api-access-dk2kd" (OuterVolumeSpecName: "kube-api-access-dk2kd") pod "3d4268b7-60e8-42ff-b0d2-dac7a000170e" (UID: "3d4268b7-60e8-42ff-b0d2-dac7a000170e"). InnerVolumeSpecName "kube-api-access-dk2kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.489481 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d4268b7-60e8-42ff-b0d2-dac7a000170e" (UID: "3d4268b7-60e8-42ff-b0d2-dac7a000170e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.503551 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.503576 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4268b7-60e8-42ff-b0d2-dac7a000170e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.503586 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk2kd\" (UniqueName: \"kubernetes.io/projected/3d4268b7-60e8-42ff-b0d2-dac7a000170e-kube-api-access-dk2kd\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.752422 4822 generic.go:334] "Generic (PLEG): container finished" podID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerID="ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505" exitCode=0 Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.752465 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqz8p" event={"ID":"3d4268b7-60e8-42ff-b0d2-dac7a000170e","Type":"ContainerDied","Data":"ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505"} Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.752490 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqz8p" event={"ID":"3d4268b7-60e8-42ff-b0d2-dac7a000170e","Type":"ContainerDied","Data":"0fc166611b41a78adb56b719cefa4c45c93165ce1560f1767fb8379189812c25"} Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.752510 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqz8p" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.752516 4822 scope.go:117] "RemoveContainer" containerID="ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.785178 4822 scope.go:117] "RemoveContainer" containerID="f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.785357 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pqz8p"] Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.791946 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pqz8p"] Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.820441 4822 scope.go:117] "RemoveContainer" containerID="050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.860499 4822 scope.go:117] "RemoveContainer" containerID="ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505" Nov 24 15:18:14 crc kubenswrapper[4822]: E1124 15:18:14.861025 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505\": container with ID starting with ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505 not found: ID does not exist" containerID="ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.861071 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505"} err="failed to get container status \"ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505\": rpc error: code = NotFound desc = could not find container \"ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505\": container with ID starting with ae130e021b810df0d1f56dfc01c538b888c2f73646bdd476f6fc4d4148781505 not found: ID does not exist" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.861097 4822 scope.go:117] "RemoveContainer" containerID="f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd" Nov 24 15:18:14 crc kubenswrapper[4822]: E1124 15:18:14.861410 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd\": container with ID starting with f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd not found: ID does not exist" containerID="f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.861492 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd"} err="failed to get container status \"f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd\": rpc error: code = NotFound desc = could not find container \"f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd\": container with ID starting with f9ae9ca99e0ecda3a383606a7004470d15af8c6056e6356c8123d122a73fe8bd not found: ID does not exist" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.861563 4822 scope.go:117] "RemoveContainer" containerID="050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8" Nov 24 15:18:14 crc kubenswrapper[4822]: E1124 15:18:14.861908 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8\": container with ID starting with 050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8 not found: ID does not exist" containerID="050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8" Nov 24 15:18:14 crc kubenswrapper[4822]: I1124 15:18:14.861954 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8"} err="failed to get container status \"050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8\": rpc error: code = NotFound desc = could not find container \"050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8\": container with ID starting with 050718fbfb6f98b3101ab5d5b990182adbc022318d398f5aa593f9b6057e35e8 not found: ID does not exist" Nov 24 15:18:15 crc kubenswrapper[4822]: I1124 15:18:15.723741 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" path="/var/lib/kubelet/pods/3d4268b7-60e8-42ff-b0d2-dac7a000170e/volumes" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.092865 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_389f0d9e-a5ab-4d8b-82e2-f4c90d830c42/init-config-reloader/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.270676 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_389f0d9e-a5ab-4d8b-82e2-f4c90d830c42/config-reloader/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.341132 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_389f0d9e-a5ab-4d8b-82e2-f4c90d830c42/init-config-reloader/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.369031 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_389f0d9e-a5ab-4d8b-82e2-f4c90d830c42/alertmanager/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.456818 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5cbff998cd-sdjjk_22ebfa19-b296-49c5-888b-cadf1e18d897/barbican-api/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.541673 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5cbff998cd-sdjjk_22ebfa19-b296-49c5-888b-cadf1e18d897/barbican-api-log/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.578774 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75cb666446-dhnfk_045d67b1-30d7-432f-a37e-165ca4eef3b4/barbican-keystone-listener/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.660472 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75cb666446-dhnfk_045d67b1-30d7-432f-a37e-165ca4eef3b4/barbican-keystone-listener-log/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.782064 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-74556c64c7-5cqch_75962c4e-af47-4216-99b1-162e00816a30/barbican-worker/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.883972 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-74556c64c7-5cqch_75962c4e-af47-4216-99b1-162e00816a30/barbican-worker-log/0.log" Nov 24 15:18:34 crc kubenswrapper[4822]: I1124 15:18:34.991274 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-wgnfk_3cea7046-1628-413a-99d0-aa36783ec7fc/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.138098 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7891e240-53f5-434a-9ee1-8e3362ac4686/ceilometer-notification-agent/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.153138 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7891e240-53f5-434a-9ee1-8e3362ac4686/ceilometer-central-agent/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.197110 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7891e240-53f5-434a-9ee1-8e3362ac4686/sg-core/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.211910 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7891e240-53f5-434a-9ee1-8e3362ac4686/proxy-httpd/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.354361 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5c756c92-6b13-4f1d-b3c0-b217c926a2c8/cinder-api-log/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.378417 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5c756c92-6b13-4f1d-b3c0-b217c926a2c8/cinder-api/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.568116 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_96b0c116-ea8d-4c76-b95d-0b95131da043/cinder-scheduler/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.622898 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_96b0c116-ea8d-4c76-b95d-0b95131da043/probe/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.701542 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_4089c54f-b1cf-40c4-8369-872dabeb8202/cloudkitty-api/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.754783 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_4089c54f-b1cf-40c4-8369-872dabeb8202/cloudkitty-api-log/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.795624 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_ac81e91d-a816-4a86-98b6-3e43f36c7b6f/loki-compactor/0.log" Nov 24 15:18:35 crc kubenswrapper[4822]: I1124 15:18:35.944136 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-56cd74f89f-mvpjv_64db7204-1d9c-48c9-92b7-1fb1d468aa6c/loki-distributor/0.log" Nov 24 15:18:36 crc kubenswrapper[4822]: I1124 15:18:36.280421 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-mz99v_b9bc089e-4276-4b2d-b7b3-8f455629074c/gateway/0.log" Nov 24 15:18:36 crc kubenswrapper[4822]: I1124 15:18:36.291305 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-s2zzq_b1986632-6691-4479-a0b5-a7dc71eda136/gateway/0.log" Nov 24 15:18:36 crc kubenswrapper[4822]: I1124 15:18:36.456964 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_8f57fdf8-920a-4329-a0b1-89b62cfa82a6/loki-index-gateway/0.log" Nov 24 15:18:36 crc kubenswrapper[4822]: I1124 15:18:36.569344 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_849f7f3c-3605-4e1b-8044-5e063841745e/loki-ingester/0.log" Nov 24 15:18:36 crc kubenswrapper[4822]: I1124 15:18:36.655971 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-548665d79b-fjkbp_f03659dc-ab31-4c65-90d4-a94c4044b76e/loki-querier/0.log" Nov 24 15:18:36 crc kubenswrapper[4822]: I1124 15:18:36.751747 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-779849886d-f557q_66287d73-c11d-48ba-aba2-570671d308e2/loki-query-frontend/0.log" Nov 24 15:18:36 crc kubenswrapper[4822]: I1124 15:18:36.949746 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-x8khj_dfb4aa15-44b2-45f7-b4a0-2b850bffa95f/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:37 crc kubenswrapper[4822]: I1124 15:18:37.178115 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-6zprf_8e390d8b-af25-4297-9ca2-e8fe094f7f71/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:37 crc kubenswrapper[4822]: I1124 15:18:37.382112 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79c5f76cb7-gj65d_62305997-e4f9-4e0e-9e28-0dea4382bf40/init/0.log" Nov 24 15:18:37 crc kubenswrapper[4822]: I1124 15:18:37.564843 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79c5f76cb7-gj65d_62305997-e4f9-4e0e-9e28-0dea4382bf40/init/0.log" Nov 24 15:18:37 crc kubenswrapper[4822]: I1124 15:18:37.625921 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79c5f76cb7-gj65d_62305997-e4f9-4e0e-9e28-0dea4382bf40/dnsmasq-dns/0.log" Nov 24 15:18:37 crc kubenswrapper[4822]: I1124 15:18:37.827962 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-f8fz5_9700df2b-526c-44d1-ba51-f9313739f6a8/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:37 crc kubenswrapper[4822]: I1124 15:18:37.874729 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9/glance-httpd/0.log" Nov 24 15:18:37 crc kubenswrapper[4822]: I1124 15:18:37.915281 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e0e7dc25-3ad1-4deb-88a0-5522ae9e6da9/glance-log/0.log" Nov 24 15:18:38 crc kubenswrapper[4822]: I1124 15:18:38.122931 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_33a75f4c-1afc-43a0-81cb-2662d187cde4/glance-log/0.log" Nov 24 15:18:38 crc kubenswrapper[4822]: I1124 15:18:38.138689 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_33a75f4c-1afc-43a0-81cb-2662d187cde4/glance-httpd/0.log" Nov 24 15:18:38 crc kubenswrapper[4822]: I1124 15:18:38.298608 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-sttpl_1eca33a0-77b5-42a6-9846-ecb6a0582a87/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:38 crc kubenswrapper[4822]: I1124 15:18:38.402441 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-pjjnf_9c88d30d-9d50-4018-8ea3-aa480b757eee/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:38 crc kubenswrapper[4822]: I1124 15:18:38.603350 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399941-rxk5k_07506623-c231-4474-9fbf-64be63010543/keystone-cron/0.log" Nov 24 15:18:38 crc kubenswrapper[4822]: I1124 15:18:38.674304 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7f958c4db4-6hhpg_d0f0a9aa-0b20-447b-9a08-9d5525010c40/keystone-api/0.log" Nov 24 15:18:38 crc kubenswrapper[4822]: I1124 15:18:38.789895 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_e3e633d9-cd8a-4366-a09a-34c653b3275c/cloudkitty-proc/0.log" Nov 24 15:18:38 crc kubenswrapper[4822]: I1124 15:18:38.817470 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3f75805d-90a9-4782-b65f-8ca7d1b91338/kube-state-metrics/0.log" Nov 24 15:18:38 crc kubenswrapper[4822]: I1124 15:18:38.905007 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-8wcxn_4b034bf8-6635-477d-a31f-b7fdf270d374/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:39 crc kubenswrapper[4822]: I1124 15:18:39.161002 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cdd657c7c-xmwlh_ef13e0bb-79cf-492c-880b-22a79fc0045b/neutron-api/0.log" Nov 24 15:18:39 crc kubenswrapper[4822]: I1124 15:18:39.213787 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cdd657c7c-xmwlh_ef13e0bb-79cf-492c-880b-22a79fc0045b/neutron-httpd/0.log" Nov 24 15:18:39 crc kubenswrapper[4822]: I1124 15:18:39.339264 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-45xbc_f6f36194-152e-41f6-9fa9-fd0f7b279843/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:39 crc kubenswrapper[4822]: I1124 15:18:39.573980 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_81b6cb5a-5c0f-49b8-ae7f-c3f81f553381/nova-api-log/0.log" Nov 24 15:18:39 crc kubenswrapper[4822]: I1124 15:18:39.792516 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3ec086ba-ec9d-4267-a5e4-0beac4c89184/nova-cell0-conductor-conductor/0.log" Nov 24 15:18:39 crc kubenswrapper[4822]: I1124 15:18:39.822098 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_81b6cb5a-5c0f-49b8-ae7f-c3f81f553381/nova-api-api/0.log" Nov 24 15:18:39 crc kubenswrapper[4822]: I1124 15:18:39.895936 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b2336f65-386b-401e-b723-abf4666c08e4/nova-cell1-conductor-conductor/0.log" Nov 24 15:18:40 crc kubenswrapper[4822]: I1124 15:18:40.069077 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_225bbbb8-99a0-4128-a39c-dc69b59eda7b/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 15:18:40 crc kubenswrapper[4822]: I1124 15:18:40.204474 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-cfwv9_cb850024-1cff-4a61-8d54-c0eff343bd4b/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:40 crc kubenswrapper[4822]: I1124 15:18:40.431136 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7d9bde3f-20ae-4497-8ce6-af4047c33364/nova-metadata-log/0.log" Nov 24 15:18:40 crc kubenswrapper[4822]: I1124 15:18:40.639332 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_fa9ec181-a56f-4d25-925e-c5b9cf962cdf/nova-scheduler-scheduler/0.log" Nov 24 15:18:40 crc kubenswrapper[4822]: I1124 15:18:40.661797 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4c617427-bd54-45c7-8e20-4ed74395e58c/mysql-bootstrap/0.log" Nov 24 15:18:40 crc kubenswrapper[4822]: I1124 15:18:40.822346 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4c617427-bd54-45c7-8e20-4ed74395e58c/mysql-bootstrap/0.log" Nov 24 15:18:40 crc kubenswrapper[4822]: I1124 15:18:40.891482 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4c617427-bd54-45c7-8e20-4ed74395e58c/galera/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.025814 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a756bbea-7fc2-443e-8d92-37f993ad9ad8/mysql-bootstrap/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.205867 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a756bbea-7fc2-443e-8d92-37f993ad9ad8/mysql-bootstrap/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.301798 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a756bbea-7fc2-443e-8d92-37f993ad9ad8/galera/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.444494 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_15618ea1-3097-4bd0-aab1-8feedfc75839/openstackclient/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.506952 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7d9bde3f-20ae-4497-8ce6-af4047c33364/nova-metadata-metadata/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.577436 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.577496 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.593095 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-n9zjj_e6c5c30b-4d09-4d55-b483-a6b1a6cfee86/openstack-network-exporter/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.750491 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7l6s7_fc49c420-99e9-498c-8b80-c7d2b2c9ed18/ovsdb-server-init/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.938739 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7l6s7_fc49c420-99e9-498c-8b80-c7d2b2c9ed18/ovs-vswitchd/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.942046 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7l6s7_fc49c420-99e9-498c-8b80-c7d2b2c9ed18/ovsdb-server-init/0.log" Nov 24 15:18:41 crc kubenswrapper[4822]: I1124 15:18:41.972339 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7l6s7_fc49c420-99e9-498c-8b80-c7d2b2c9ed18/ovsdb-server/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.159493 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vh28h_bb7d7a89-bc89-4b18-bbf2-626afb9452b3/ovn-controller/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.178106 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-jxppt_b90073ee-ca05-4991-a028-96fa152f8a71/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.390928 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a/ovn-northd/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.444434 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a1e7b65a-52b5-4f4b-aa89-f6247e4fa31a/openstack-network-exporter/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.633432 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_890c0404-624f-4c2c-a8d8-ca9aad10c7af/ovsdbserver-nb/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.693352 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_890c0404-624f-4c2c-a8d8-ca9aad10c7af/openstack-network-exporter/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.732149 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_830f9ad0-3965-4887-b131-df42fdb35db6/openstack-network-exporter/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.858301 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_830f9ad0-3965-4887-b131-df42fdb35db6/ovsdbserver-sb/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.985018 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d6f884c6d-xlmbh_0eb98c2c-948c-4ed6-99b6-f6b352d6d50c/placement-api/0.log" Nov 24 15:18:42 crc kubenswrapper[4822]: I1124 15:18:42.990732 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d6f884c6d-xlmbh_0eb98c2c-948c-4ed6-99b6-f6b352d6d50c/placement-log/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.053670 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_51214ce1-190b-4166-9817-af60ba07c73e/init-config-reloader/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.241850 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_51214ce1-190b-4166-9817-af60ba07c73e/config-reloader/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.290155 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_51214ce1-190b-4166-9817-af60ba07c73e/prometheus/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.315654 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_51214ce1-190b-4166-9817-af60ba07c73e/init-config-reloader/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.393745 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_51214ce1-190b-4166-9817-af60ba07c73e/thanos-sidecar/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.532886 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c969e517-0251-4537-9999-53bf0cfa7cf0/setup-container/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.737641 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c969e517-0251-4537-9999-53bf0cfa7cf0/setup-container/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.790404 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c969e517-0251-4537-9999-53bf0cfa7cf0/rabbitmq/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.794631 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_af242c06-68a2-436b-84d9-c2d020f15903/setup-container/0.log" Nov 24 15:18:43 crc kubenswrapper[4822]: I1124 15:18:43.986113 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_af242c06-68a2-436b-84d9-c2d020f15903/rabbitmq/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.021516 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_af242c06-68a2-436b-84d9-c2d020f15903/setup-container/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.076345 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-6gk5s_ea1150ac-561e-4776-a2b4-233a5d4189be/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.222953 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-dx95w_0fc7b81a-075b-4f00-b4b1-d9863317c650/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.245439 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lvn5d_4b17a1c2-3060-401f-89a2-21f6e7edec54/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.418318 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-xkbkk_a385f6bf-f192-408b-9ed4-fe18f7784682/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.530497 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-5ldtr_3e26bb17-0138-493a-90d6-8164ca8367dd/ssh-known-hosts-edpm-deployment/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.736450 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6b8fb4f557-bkwxl_753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf/proxy-server/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.752679 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6b8fb4f557-bkwxl_753b9eec-c5d7-4cbd-8a0d-e1724a6c5ccf/proxy-httpd/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.890875 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-wx524_65b23262-b92d-4338-8e81-7e594f48be7b/swift-ring-rebalance/0.log" Nov 24 15:18:44 crc kubenswrapper[4822]: I1124 15:18:44.970454 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/account-auditor/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.051523 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/account-reaper/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.057595 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/account-replicator/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.479943 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/account-server/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.496041 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/container-replicator/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.504759 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/container-auditor/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.546840 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/container-server/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.711764 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/object-auditor/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.781291 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/container-updater/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.820465 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/object-expirer/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.835737 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/object-replicator/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.936680 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/object-server/0.log" Nov 24 15:18:45 crc kubenswrapper[4822]: I1124 15:18:45.991402 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/object-updater/0.log" Nov 24 15:18:46 crc kubenswrapper[4822]: I1124 15:18:46.013650 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/rsync/0.log" Nov 24 15:18:46 crc kubenswrapper[4822]: I1124 15:18:46.054327 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4acc7e6a-472b-468a-b709-183f8b3c2b5b/swift-recon-cron/0.log" Nov 24 15:18:46 crc kubenswrapper[4822]: I1124 15:18:46.248905 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s22zg_561e27e7-d8ea-4764-872e-0b022659275c/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:46 crc kubenswrapper[4822]: I1124 15:18:46.318426 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-7xv4k_3b528ac3-bde2-404d-b02b-9e7a56710af4/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 15:18:53 crc kubenswrapper[4822]: I1124 15:18:53.237003 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_346023fa-f08b-4ddb-a527-a637be223953/memcached/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.180647 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-pvs2h_e260fa6b-7576-4e48-8025-2c7ef2bd97f7/kube-rbac-proxy/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.247656 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-pvs2h_e260fa6b-7576-4e48-8025-2c7ef2bd97f7/manager/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.337586 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-6f4n4_10f3ea0d-a23e-47eb-b344-0696f8c9b1e8/kube-rbac-proxy/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.387721 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-6f4n4_10f3ea0d-a23e-47eb-b344-0696f8c9b1e8/manager/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.506535 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-tvqv6_3809d742-f676-41dd-aae2-d03be501e0b0/kube-rbac-proxy/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.534590 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-tvqv6_3809d742-f676-41dd-aae2-d03be501e0b0/manager/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.577745 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.577812 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.645589 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw_2f49b7d5-56b4-4c58-ba64-866a06860c5a/util/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.771663 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw_2f49b7d5-56b4-4c58-ba64-866a06860c5a/util/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.775801 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw_2f49b7d5-56b4-4c58-ba64-866a06860c5a/pull/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.827942 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw_2f49b7d5-56b4-4c58-ba64-866a06860c5a/pull/0.log" Nov 24 15:19:11 crc kubenswrapper[4822]: I1124 15:19:11.995722 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw_2f49b7d5-56b4-4c58-ba64-866a06860c5a/pull/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.020452 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw_2f49b7d5-56b4-4c58-ba64-866a06860c5a/extract/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.024814 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f6abe9d955c8b0019ff9d2fb85f18ddb453ac410a764cabb309dbef48frn4kw_2f49b7d5-56b4-4c58-ba64-866a06860c5a/util/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.215911 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-xh9nm_70ff7dde-abd4-453c-8164-c7304292c41d/kube-rbac-proxy/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.237849 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-xh9nm_70ff7dde-abd4-453c-8164-c7304292c41d/manager/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.282258 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-k6hqr_57aa63d3-2464-4cc7-a5e1-539e60ee6cca/kube-rbac-proxy/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.387035 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-k6hqr_57aa63d3-2464-4cc7-a5e1-539e60ee6cca/manager/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.463865 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-4fgsn_5047b32c-bd21-4f83-bce1-248a1ce109e9/kube-rbac-proxy/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.487560 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-4fgsn_5047b32c-bd21-4f83-bce1-248a1ce109e9/manager/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.599502 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-9j9ls_37a54972-a29e-4ae1-a3f1-104bc634b3cc/kube-rbac-proxy/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.768130 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-z8mj6_a22d61a1-a9c4-47e9-8c5d-6593642e798b/kube-rbac-proxy/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.775898 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-9j9ls_37a54972-a29e-4ae1-a3f1-104bc634b3cc/manager/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.779379 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-z8mj6_a22d61a1-a9c4-47e9-8c5d-6593642e798b/manager/0.log" Nov 24 15:19:12 crc kubenswrapper[4822]: I1124 15:19:12.979449 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-2vr5g_ac525cf9-4bbf-4030-a7b3-f2da129d81c8/kube-rbac-proxy/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.008875 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-2vr5g_ac525cf9-4bbf-4030-a7b3-f2da129d81c8/manager/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.103313 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-khnxm_8939f780-02b5-4022-80cb-a6fd247aa7f8/kube-rbac-proxy/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.150793 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-khnxm_8939f780-02b5-4022-80cb-a6fd247aa7f8/manager/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.198836 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-jkq8m_3e001a57-d6f1-4d8a-9d51-f347063a55b5/kube-rbac-proxy/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.283483 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-jkq8m_3e001a57-d6f1-4d8a-9d51-f347063a55b5/manager/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.340524 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-bzj92_80740002-5d02-4aa7-b6fa-3de642266f21/kube-rbac-proxy/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.467792 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-bzj92_80740002-5d02-4aa7-b6fa-3de642266f21/manager/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.600526 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-xd82t_fd0cee72-09cc-449b-a6f7-fb951f0b87ba/manager/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.628156 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-xd82t_fd0cee72-09cc-449b-a6f7-fb951f0b87ba/kube-rbac-proxy/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.727698 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-mf4hq_b5737c88-eb91-432d-a620-734337051260/kube-rbac-proxy/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.763895 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-mf4hq_b5737c88-eb91-432d-a620-734337051260/manager/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.902965 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc_b01f0f96-a40f-407a-a610-bd2e945a3e49/manager/0.log" Nov 24 15:19:13 crc kubenswrapper[4822]: I1124 15:19:13.950101 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-tlhrc_b01f0f96-a40f-407a-a610-bd2e945a3e49/kube-rbac-proxy/0.log" Nov 24 15:19:14 crc kubenswrapper[4822]: I1124 15:19:14.074835 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7b58687767-tfthk_a648d53c-be72-46ed-8308-79565a88bd4c/kube-rbac-proxy/0.log" Nov 24 15:19:14 crc kubenswrapper[4822]: I1124 15:19:14.171143 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-68f4d784d9-89pr8_369ebe2a-4e3b-4131-8d92-8f8066f2fd78/kube-rbac-proxy/0.log" Nov 24 15:19:14 crc kubenswrapper[4822]: I1124 15:19:14.397521 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-527lx_1899e20a-040e-4111-a386-e23e95850d2c/registry-server/0.log" Nov 24 15:19:14 crc kubenswrapper[4822]: I1124 15:19:14.527693 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-68f4d784d9-89pr8_369ebe2a-4e3b-4131-8d92-8f8066f2fd78/operator/0.log" Nov 24 15:19:14 crc kubenswrapper[4822]: I1124 15:19:14.599147 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-lsbgj_dbffd719-e3e2-4728-9301-da9efe8eb57b/kube-rbac-proxy/0.log" Nov 24 15:19:14 crc kubenswrapper[4822]: I1124 15:19:14.733849 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-lsbgj_dbffd719-e3e2-4728-9301-da9efe8eb57b/manager/0.log" Nov 24 15:19:14 crc kubenswrapper[4822]: I1124 15:19:14.824069 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-45wdc_d758b1f7-956b-4881-8a61-a81e55fdffcd/manager/0.log" Nov 24 15:19:14 crc kubenswrapper[4822]: I1124 15:19:14.835711 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-45wdc_d758b1f7-956b-4881-8a61-a81e55fdffcd/kube-rbac-proxy/0.log" Nov 24 15:19:14 crc kubenswrapper[4822]: I1124 15:19:14.975468 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-98xcn_4e5bcbaf-c685-4047-a37a-3c058f8e1cbc/operator/0.log" Nov 24 15:19:15 crc kubenswrapper[4822]: I1124 15:19:15.016379 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-wtv4t_38159b16-c3af-4bcd-8734-88cfef861663/kube-rbac-proxy/0.log" Nov 24 15:19:15 crc kubenswrapper[4822]: I1124 15:19:15.155626 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7b58687767-tfthk_a648d53c-be72-46ed-8308-79565a88bd4c/manager/0.log" Nov 24 15:19:15 crc kubenswrapper[4822]: I1124 15:19:15.216580 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-54d7678447-dx47k_253630d1-2b9c-4bc4-a7fd-357c66d00f9e/kube-rbac-proxy/0.log" Nov 24 15:19:15 crc kubenswrapper[4822]: I1124 15:19:15.222827 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-wtv4t_38159b16-c3af-4bcd-8734-88cfef861663/manager/0.log" Nov 24 15:19:15 crc kubenswrapper[4822]: I1124 15:19:15.452892 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-2p2qp_97913a25-5fcf-4541-b928-191ffe821e93/manager/0.log" Nov 24 15:19:15 crc kubenswrapper[4822]: I1124 15:19:15.455033 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-2p2qp_97913a25-5fcf-4541-b928-191ffe821e93/kube-rbac-proxy/0.log" Nov 24 15:19:15 crc kubenswrapper[4822]: I1124 15:19:15.517831 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-54d7678447-dx47k_253630d1-2b9c-4bc4-a7fd-357c66d00f9e/manager/0.log" Nov 24 15:19:15 crc kubenswrapper[4822]: I1124 15:19:15.558389 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-dc279_980ad0a7-70cd-4f77-bc29-119868034087/kube-rbac-proxy/0.log" Nov 24 15:19:15 crc kubenswrapper[4822]: I1124 15:19:15.621310 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-dc279_980ad0a7-70cd-4f77-bc29-119868034087/manager/0.log" Nov 24 15:19:32 crc kubenswrapper[4822]: I1124 15:19:32.317687 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-vqxq9_6b8514e2-4241-4916-b843-d62897d4af2c/control-plane-machine-set-operator/0.log" Nov 24 15:19:32 crc kubenswrapper[4822]: I1124 15:19:32.497384 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qr4nb_2446a52a-a882-4ebd-972d-eb61d2e6c086/machine-api-operator/0.log" Nov 24 15:19:32 crc kubenswrapper[4822]: I1124 15:19:32.513908 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qr4nb_2446a52a-a882-4ebd-972d-eb61d2e6c086/kube-rbac-proxy/0.log" Nov 24 15:19:41 crc kubenswrapper[4822]: I1124 15:19:41.577954 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:19:41 crc kubenswrapper[4822]: I1124 15:19:41.578769 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:19:41 crc kubenswrapper[4822]: I1124 15:19:41.578835 4822 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nst99" Nov 24 15:19:41 crc kubenswrapper[4822]: I1124 15:19:41.579998 4822 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5"} pod="openshift-machine-config-operator/machine-config-daemon-nst99" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:19:41 crc kubenswrapper[4822]: I1124 15:19:41.580098 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" containerID="cri-o://e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" gracePeriod=600 Nov 24 15:19:41 crc kubenswrapper[4822]: E1124 15:19:41.703686 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:19:42 crc kubenswrapper[4822]: I1124 15:19:42.637528 4822 generic.go:334] "Generic (PLEG): container finished" podID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" exitCode=0 Nov 24 15:19:42 crc kubenswrapper[4822]: I1124 15:19:42.637568 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerDied","Data":"e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5"} Nov 24 15:19:42 crc kubenswrapper[4822]: I1124 15:19:42.637623 4822 scope.go:117] "RemoveContainer" containerID="7b89a251fd0f0cfe8d88737b003ae9166ecdc96e5233f626b16b042314ed274c" Nov 24 15:19:42 crc kubenswrapper[4822]: I1124 15:19:42.638406 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:19:42 crc kubenswrapper[4822]: E1124 15:19:42.638813 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:19:46 crc kubenswrapper[4822]: I1124 15:19:46.729502 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-jnzqf_e73af35b-1628-45a3-8a96-1e20b8bca297/cert-manager-controller/0.log" Nov 24 15:19:46 crc kubenswrapper[4822]: I1124 15:19:46.977848 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-drv4n_d97bac64-5623-41f7-8338-19ad1704e5de/cert-manager-cainjector/0.log" Nov 24 15:19:46 crc kubenswrapper[4822]: I1124 15:19:46.985614 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-hxldz_640c8f8a-b96f-49ba-afd5-c4c80ac909c0/cert-manager-webhook/0.log" Nov 24 15:19:55 crc kubenswrapper[4822]: I1124 15:19:55.705814 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:19:55 crc kubenswrapper[4822]: E1124 15:19:55.706773 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:20:01 crc kubenswrapper[4822]: I1124 15:20:01.502837 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-96lpb_4dc5455a-d1ab-40b6-942f-00e55730b89b/nmstate-console-plugin/0.log" Nov 24 15:20:01 crc kubenswrapper[4822]: I1124 15:20:01.661685 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-2ns9q_1e2f143c-b018-4732-b91d-bfecee8726a9/nmstate-metrics/0.log" Nov 24 15:20:01 crc kubenswrapper[4822]: I1124 15:20:01.695627 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-2ns9q_1e2f143c-b018-4732-b91d-bfecee8726a9/kube-rbac-proxy/0.log" Nov 24 15:20:01 crc kubenswrapper[4822]: I1124 15:20:01.715066 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-xljmw_27e6c538-998e-46e2-acba-eb9eb622910f/nmstate-handler/0.log" Nov 24 15:20:01 crc kubenswrapper[4822]: I1124 15:20:01.956474 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-hnmjh_08639b4b-1faa-4397-9727-b752d99978bc/nmstate-operator/0.log" Nov 24 15:20:02 crc kubenswrapper[4822]: I1124 15:20:02.077351 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-zgvgl_56484cc3-efd5-4514-927c-350149fa1362/nmstate-webhook/0.log" Nov 24 15:20:07 crc kubenswrapper[4822]: I1124 15:20:07.705261 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:20:07 crc kubenswrapper[4822]: E1124 15:20:07.706143 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:20:15 crc kubenswrapper[4822]: I1124 15:20:15.880842 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-db9f8c7d4-kz5dv_cef545df-04fd-49bd-b98a-bcb23fd874f0/kube-rbac-proxy/0.log" Nov 24 15:20:15 crc kubenswrapper[4822]: I1124 15:20:15.917745 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-db9f8c7d4-kz5dv_cef545df-04fd-49bd-b98a-bcb23fd874f0/manager/0.log" Nov 24 15:20:18 crc kubenswrapper[4822]: I1124 15:20:18.704244 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:20:18 crc kubenswrapper[4822]: E1124 15:20:18.704939 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:20:29 crc kubenswrapper[4822]: I1124 15:20:29.705382 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:20:29 crc kubenswrapper[4822]: E1124 15:20:29.706043 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.130075 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-dlvwp_ed1304c2-b03d-45de-a7c2-dfc08da3691d/kube-rbac-proxy/0.log" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.239525 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-dlvwp_ed1304c2-b03d-45de-a7c2-dfc08da3691d/controller/0.log" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.366038 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-frr-files/0.log" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.565726 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-metrics/0.log" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.565935 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-reloader/0.log" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.576721 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-frr-files/0.log" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.591685 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-reloader/0.log" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.911670 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-frr-files/0.log" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.917879 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-reloader/0.log" Nov 24 15:20:30 crc kubenswrapper[4822]: I1124 15:20:30.984686 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-metrics/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.008550 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-metrics/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.085182 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-frr-files/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.100559 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-reloader/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.163911 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/cp-metrics/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.195966 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/controller/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.316762 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/frr-metrics/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.392403 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/kube-rbac-proxy/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.413387 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/kube-rbac-proxy-frr/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.567195 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/reloader/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.623651 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-9ld68_7144b776-5528-4a9a-9489-6e73c2608485/frr-k8s-webhook-server/0.log" Nov 24 15:20:31 crc kubenswrapper[4822]: I1124 15:20:31.872865 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-85fcb85fb9-hk6dw_266b3b36-bac6-482f-b35e-5153d0c3b099/manager/0.log" Nov 24 15:20:32 crc kubenswrapper[4822]: I1124 15:20:32.068680 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5fbf9ff7c-xsgkp_0b14b176-790c-4e60-9f93-bda90fda768a/webhook-server/0.log" Nov 24 15:20:32 crc kubenswrapper[4822]: I1124 15:20:32.090249 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qd7vl_2d55bf7d-877f-4605-9eea-b792b465e271/kube-rbac-proxy/0.log" Nov 24 15:20:32 crc kubenswrapper[4822]: I1124 15:20:32.637070 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ts9cq_6ca85134-1017-4276-85a1-ac5cf4c47c0b/frr/0.log" Nov 24 15:20:32 crc kubenswrapper[4822]: I1124 15:20:32.719822 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qd7vl_2d55bf7d-877f-4605-9eea-b792b465e271/speaker/0.log" Nov 24 15:20:40 crc kubenswrapper[4822]: I1124 15:20:40.704182 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:20:40 crc kubenswrapper[4822]: E1124 15:20:40.704948 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:20:47 crc kubenswrapper[4822]: I1124 15:20:47.567499 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5_20e971ff-ea2d-4652-901b-f744c846972b/util/0.log" Nov 24 15:20:47 crc kubenswrapper[4822]: I1124 15:20:47.693094 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5_20e971ff-ea2d-4652-901b-f744c846972b/util/0.log" Nov 24 15:20:47 crc kubenswrapper[4822]: I1124 15:20:47.714930 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5_20e971ff-ea2d-4652-901b-f744c846972b/pull/0.log" Nov 24 15:20:47 crc kubenswrapper[4822]: I1124 15:20:47.775357 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5_20e971ff-ea2d-4652-901b-f744c846972b/pull/0.log" Nov 24 15:20:47 crc kubenswrapper[4822]: I1124 15:20:47.932801 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5_20e971ff-ea2d-4652-901b-f744c846972b/util/0.log" Nov 24 15:20:47 crc kubenswrapper[4822]: I1124 15:20:47.933996 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5_20e971ff-ea2d-4652-901b-f744c846972b/extract/0.log" Nov 24 15:20:47 crc kubenswrapper[4822]: I1124 15:20:47.944139 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694bdrn5_20e971ff-ea2d-4652-901b-f744c846972b/pull/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.105560 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx_7f24b1dd-5e33-4c52-af2d-43a4c9849225/util/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.233290 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx_7f24b1dd-5e33-4c52-af2d-43a4c9849225/util/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.251836 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx_7f24b1dd-5e33-4c52-af2d-43a4c9849225/pull/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.261049 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx_7f24b1dd-5e33-4c52-af2d-43a4c9849225/pull/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.461683 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx_7f24b1dd-5e33-4c52-af2d-43a4c9849225/extract/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.487713 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx_7f24b1dd-5e33-4c52-af2d-43a4c9849225/pull/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.492917 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e3039q5vx_7f24b1dd-5e33-4c52-af2d-43a4c9849225/util/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.621764 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8_1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0/util/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.791881 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8_1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0/pull/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.818224 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8_1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0/util/0.log" Nov 24 15:20:48 crc kubenswrapper[4822]: I1124 15:20:48.847428 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8_1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0/pull/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.039406 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8_1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0/pull/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.042629 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8_1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0/extract/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.082917 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e262q8_1b8609cb-14ea-4cb2-8ae5-5b42a59a11d0/util/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.204232 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l_a69ba74b-663d-4777-b50e-6988f9a9a894/util/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.373266 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l_a69ba74b-663d-4777-b50e-6988f9a9a894/util/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.419968 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l_a69ba74b-663d-4777-b50e-6988f9a9a894/pull/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.419975 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l_a69ba74b-663d-4777-b50e-6988f9a9a894/pull/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.575684 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l_a69ba74b-663d-4777-b50e-6988f9a9a894/pull/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.584338 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l_a69ba74b-663d-4777-b50e-6988f9a9a894/extract/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.619744 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dkj8l_a69ba74b-663d-4777-b50e-6988f9a9a894/util/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.759600 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-k4tjc_189cb4b9-5c76-4872-b350-29525973b831/extract-utilities/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.938691 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-k4tjc_189cb4b9-5c76-4872-b350-29525973b831/extract-content/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.941229 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-k4tjc_189cb4b9-5c76-4872-b350-29525973b831/extract-utilities/0.log" Nov 24 15:20:49 crc kubenswrapper[4822]: I1124 15:20:49.987018 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-k4tjc_189cb4b9-5c76-4872-b350-29525973b831/extract-content/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.067597 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-k4tjc_189cb4b9-5c76-4872-b350-29525973b831/extract-utilities/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.111741 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-k4tjc_189cb4b9-5c76-4872-b350-29525973b831/extract-content/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.145079 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-k4tjc_189cb4b9-5c76-4872-b350-29525973b831/registry-server/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.288317 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tq6m5_fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e/extract-utilities/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.474969 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tq6m5_fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e/extract-content/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.491190 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tq6m5_fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e/extract-content/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.519789 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tq6m5_fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e/extract-utilities/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.696849 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tq6m5_fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e/extract-utilities/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.719379 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94_cacd21c8-e424-492a-ba96-b618800b3d58/util/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.723027 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tq6m5_fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e/extract-content/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.975946 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94_cacd21c8-e424-492a-ba96-b618800b3d58/util/0.log" Nov 24 15:20:50 crc kubenswrapper[4822]: I1124 15:20:50.997673 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94_cacd21c8-e424-492a-ba96-b618800b3d58/pull/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.013172 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94_cacd21c8-e424-492a-ba96-b618800b3d58/pull/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.027799 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tq6m5_fa4a0ef1-2f1d-4588-b26e-9e3251de7f5e/registry-server/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.155780 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94_cacd21c8-e424-492a-ba96-b618800b3d58/pull/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.176431 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94_cacd21c8-e424-492a-ba96-b618800b3d58/util/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.176576 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6h9f94_cacd21c8-e424-492a-ba96-b618800b3d58/extract/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.209479 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xndhl_1d96a94a-717c-42c8-90e2-ae98a6edcb99/marketplace-operator/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.325853 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vlfkr_e5fca0b7-7cab-4a35-a0ef-88a75d92707a/extract-utilities/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.506077 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vlfkr_e5fca0b7-7cab-4a35-a0ef-88a75d92707a/extract-utilities/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.506081 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vlfkr_e5fca0b7-7cab-4a35-a0ef-88a75d92707a/extract-content/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.512798 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vlfkr_e5fca0b7-7cab-4a35-a0ef-88a75d92707a/extract-content/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.695634 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vlfkr_e5fca0b7-7cab-4a35-a0ef-88a75d92707a/extract-utilities/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.752622 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nvv4m_39db4f3b-fa4e-41ad-92c9-5e682207c429/extract-utilities/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.765880 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vlfkr_e5fca0b7-7cab-4a35-a0ef-88a75d92707a/extract-content/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.868693 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vlfkr_e5fca0b7-7cab-4a35-a0ef-88a75d92707a/registry-server/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.946232 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nvv4m_39db4f3b-fa4e-41ad-92c9-5e682207c429/extract-utilities/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.950360 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nvv4m_39db4f3b-fa4e-41ad-92c9-5e682207c429/extract-content/0.log" Nov 24 15:20:51 crc kubenswrapper[4822]: I1124 15:20:51.984265 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nvv4m_39db4f3b-fa4e-41ad-92c9-5e682207c429/extract-content/0.log" Nov 24 15:20:52 crc kubenswrapper[4822]: I1124 15:20:52.124751 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nvv4m_39db4f3b-fa4e-41ad-92c9-5e682207c429/extract-utilities/0.log" Nov 24 15:20:52 crc kubenswrapper[4822]: I1124 15:20:52.181112 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nvv4m_39db4f3b-fa4e-41ad-92c9-5e682207c429/extract-content/0.log" Nov 24 15:20:52 crc kubenswrapper[4822]: I1124 15:20:52.521891 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nvv4m_39db4f3b-fa4e-41ad-92c9-5e682207c429/registry-server/0.log" Nov 24 15:20:54 crc kubenswrapper[4822]: I1124 15:20:54.712762 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:20:54 crc kubenswrapper[4822]: E1124 15:20:54.714073 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:21:05 crc kubenswrapper[4822]: I1124 15:21:05.111208 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-ql56t_387c7641-9e5d-4ea0-a86f-c890d68d2903/prometheus-operator/0.log" Nov 24 15:21:05 crc kubenswrapper[4822]: I1124 15:21:05.223096 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b6d578495-tzppm_c463043b-1fc8-49c3-80a7-ac3469c1504a/prometheus-operator-admission-webhook/0.log" Nov 24 15:21:05 crc kubenswrapper[4822]: I1124 15:21:05.268197 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b6d578495-x96vd_bdf59941-e2f3-4ccf-a700-992e2d3871af/prometheus-operator-admission-webhook/0.log" Nov 24 15:21:05 crc kubenswrapper[4822]: I1124 15:21:05.407165 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-sdtbh_991aab37-0cca-4f05-a7ff-ff64d2cac3d4/operator/0.log" Nov 24 15:21:05 crc kubenswrapper[4822]: I1124 15:21:05.483865 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-8h9ks_394cbee2-1b78-4d5f-b663-317a8d4ff56e/perses-operator/0.log" Nov 24 15:21:05 crc kubenswrapper[4822]: I1124 15:21:05.704357 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:21:05 crc kubenswrapper[4822]: E1124 15:21:05.704804 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:21:18 crc kubenswrapper[4822]: I1124 15:21:18.138619 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-db9f8c7d4-kz5dv_cef545df-04fd-49bd-b98a-bcb23fd874f0/kube-rbac-proxy/0.log" Nov 24 15:21:18 crc kubenswrapper[4822]: I1124 15:21:18.157497 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-db9f8c7d4-kz5dv_cef545df-04fd-49bd-b98a-bcb23fd874f0/manager/0.log" Nov 24 15:21:19 crc kubenswrapper[4822]: I1124 15:21:19.706095 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:21:19 crc kubenswrapper[4822]: E1124 15:21:19.707157 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:21:34 crc kubenswrapper[4822]: I1124 15:21:34.704956 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:21:34 crc kubenswrapper[4822]: E1124 15:21:34.705766 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:21:45 crc kubenswrapper[4822]: I1124 15:21:45.706861 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:21:45 crc kubenswrapper[4822]: E1124 15:21:45.707801 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:21:58 crc kubenswrapper[4822]: I1124 15:21:58.704545 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:21:58 crc kubenswrapper[4822]: E1124 15:21:58.705954 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:22:09 crc kubenswrapper[4822]: I1124 15:22:09.704153 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:22:09 crc kubenswrapper[4822]: E1124 15:22:09.705140 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:22:21 crc kubenswrapper[4822]: I1124 15:22:21.704454 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:22:21 crc kubenswrapper[4822]: E1124 15:22:21.705581 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.739558 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s4c7d"] Nov 24 15:22:24 crc kubenswrapper[4822]: E1124 15:22:24.740530 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerName="registry-server" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.740543 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerName="registry-server" Nov 24 15:22:24 crc kubenswrapper[4822]: E1124 15:22:24.740564 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerName="extract-utilities" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.740570 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerName="extract-utilities" Nov 24 15:22:24 crc kubenswrapper[4822]: E1124 15:22:24.740583 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerName="extract-content" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.740589 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerName="extract-content" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.740810 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4268b7-60e8-42ff-b0d2-dac7a000170e" containerName="registry-server" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.742320 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.770252 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4c7d"] Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.821457 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-utilities\") pod \"redhat-marketplace-s4c7d\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.821728 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q4bq\" (UniqueName: \"kubernetes.io/projected/02c1c1e6-79ea-48e3-bd98-c46f7538d205-kube-api-access-5q4bq\") pod \"redhat-marketplace-s4c7d\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.822029 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-catalog-content\") pod \"redhat-marketplace-s4c7d\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.923776 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q4bq\" (UniqueName: \"kubernetes.io/projected/02c1c1e6-79ea-48e3-bd98-c46f7538d205-kube-api-access-5q4bq\") pod \"redhat-marketplace-s4c7d\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.923867 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-catalog-content\") pod \"redhat-marketplace-s4c7d\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.923973 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-utilities\") pod \"redhat-marketplace-s4c7d\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.924450 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-utilities\") pod \"redhat-marketplace-s4c7d\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.924982 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-catalog-content\") pod \"redhat-marketplace-s4c7d\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:24 crc kubenswrapper[4822]: I1124 15:22:24.953231 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q4bq\" (UniqueName: \"kubernetes.io/projected/02c1c1e6-79ea-48e3-bd98-c46f7538d205-kube-api-access-5q4bq\") pod \"redhat-marketplace-s4c7d\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:25 crc kubenswrapper[4822]: I1124 15:22:25.075168 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:25 crc kubenswrapper[4822]: I1124 15:22:25.597035 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4c7d"] Nov 24 15:22:26 crc kubenswrapper[4822]: I1124 15:22:26.244592 4822 generic.go:334] "Generic (PLEG): container finished" podID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerID="98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35" exitCode=0 Nov 24 15:22:26 crc kubenswrapper[4822]: I1124 15:22:26.244664 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4c7d" event={"ID":"02c1c1e6-79ea-48e3-bd98-c46f7538d205","Type":"ContainerDied","Data":"98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35"} Nov 24 15:22:26 crc kubenswrapper[4822]: I1124 15:22:26.245301 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4c7d" event={"ID":"02c1c1e6-79ea-48e3-bd98-c46f7538d205","Type":"ContainerStarted","Data":"9a0492cbda52adcc903e2e5d002a9f1bd598b70016123034b60b7e30b0c81b41"} Nov 24 15:22:26 crc kubenswrapper[4822]: I1124 15:22:26.246890 4822 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:22:27 crc kubenswrapper[4822]: I1124 15:22:27.253773 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4c7d" event={"ID":"02c1c1e6-79ea-48e3-bd98-c46f7538d205","Type":"ContainerStarted","Data":"d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8"} Nov 24 15:22:28 crc kubenswrapper[4822]: I1124 15:22:28.265775 4822 generic.go:334] "Generic (PLEG): container finished" podID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerID="d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8" exitCode=0 Nov 24 15:22:28 crc kubenswrapper[4822]: I1124 15:22:28.265827 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4c7d" event={"ID":"02c1c1e6-79ea-48e3-bd98-c46f7538d205","Type":"ContainerDied","Data":"d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8"} Nov 24 15:22:29 crc kubenswrapper[4822]: I1124 15:22:29.289418 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4c7d" event={"ID":"02c1c1e6-79ea-48e3-bd98-c46f7538d205","Type":"ContainerStarted","Data":"9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de"} Nov 24 15:22:29 crc kubenswrapper[4822]: I1124 15:22:29.317020 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s4c7d" podStartSLOduration=2.829827336 podStartE2EDuration="5.316967411s" podCreationTimestamp="2025-11-24 15:22:24 +0000 UTC" firstStartedPulling="2025-11-24 15:22:26.246629892 +0000 UTC m=+3783.363270379" lastFinishedPulling="2025-11-24 15:22:28.733769937 +0000 UTC m=+3785.850410454" observedRunningTime="2025-11-24 15:22:29.311493559 +0000 UTC m=+3786.428134106" watchObservedRunningTime="2025-11-24 15:22:29.316967411 +0000 UTC m=+3786.433607898" Nov 24 15:22:35 crc kubenswrapper[4822]: I1124 15:22:35.075461 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:35 crc kubenswrapper[4822]: I1124 15:22:35.077017 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:35 crc kubenswrapper[4822]: I1124 15:22:35.147092 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:35 crc kubenswrapper[4822]: I1124 15:22:35.434583 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:35 crc kubenswrapper[4822]: I1124 15:22:35.498150 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4c7d"] Nov 24 15:22:35 crc kubenswrapper[4822]: I1124 15:22:35.705647 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:22:35 crc kubenswrapper[4822]: E1124 15:22:35.706144 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:22:37 crc kubenswrapper[4822]: I1124 15:22:37.393588 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s4c7d" podUID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerName="registry-server" containerID="cri-o://9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de" gracePeriod=2 Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.083219 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.163251 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-utilities\") pod \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.163420 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-catalog-content\") pod \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.163553 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q4bq\" (UniqueName: \"kubernetes.io/projected/02c1c1e6-79ea-48e3-bd98-c46f7538d205-kube-api-access-5q4bq\") pod \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\" (UID: \"02c1c1e6-79ea-48e3-bd98-c46f7538d205\") " Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.164113 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-utilities" (OuterVolumeSpecName: "utilities") pod "02c1c1e6-79ea-48e3-bd98-c46f7538d205" (UID: "02c1c1e6-79ea-48e3-bd98-c46f7538d205"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.168812 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02c1c1e6-79ea-48e3-bd98-c46f7538d205-kube-api-access-5q4bq" (OuterVolumeSpecName: "kube-api-access-5q4bq") pod "02c1c1e6-79ea-48e3-bd98-c46f7538d205" (UID: "02c1c1e6-79ea-48e3-bd98-c46f7538d205"). InnerVolumeSpecName "kube-api-access-5q4bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.181653 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02c1c1e6-79ea-48e3-bd98-c46f7538d205" (UID: "02c1c1e6-79ea-48e3-bd98-c46f7538d205"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.266785 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.266815 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c1c1e6-79ea-48e3-bd98-c46f7538d205-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.266828 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q4bq\" (UniqueName: \"kubernetes.io/projected/02c1c1e6-79ea-48e3-bd98-c46f7538d205-kube-api-access-5q4bq\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.408094 4822 generic.go:334] "Generic (PLEG): container finished" podID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerID="9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de" exitCode=0 Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.408142 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4c7d" event={"ID":"02c1c1e6-79ea-48e3-bd98-c46f7538d205","Type":"ContainerDied","Data":"9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de"} Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.408172 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4c7d" event={"ID":"02c1c1e6-79ea-48e3-bd98-c46f7538d205","Type":"ContainerDied","Data":"9a0492cbda52adcc903e2e5d002a9f1bd598b70016123034b60b7e30b0c81b41"} Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.408191 4822 scope.go:117] "RemoveContainer" containerID="9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.408355 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4c7d" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.453335 4822 scope.go:117] "RemoveContainer" containerID="d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.454708 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4c7d"] Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.465481 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4c7d"] Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.484503 4822 scope.go:117] "RemoveContainer" containerID="98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.546926 4822 scope.go:117] "RemoveContainer" containerID="9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de" Nov 24 15:22:38 crc kubenswrapper[4822]: E1124 15:22:38.547689 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de\": container with ID starting with 9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de not found: ID does not exist" containerID="9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.547738 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de"} err="failed to get container status \"9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de\": rpc error: code = NotFound desc = could not find container \"9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de\": container with ID starting with 9be5bdd3de227b06090ff8b8ccd3268d6c4f08423cc8478d4d3e73f357f434de not found: ID does not exist" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.547770 4822 scope.go:117] "RemoveContainer" containerID="d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8" Nov 24 15:22:38 crc kubenswrapper[4822]: E1124 15:22:38.548108 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8\": container with ID starting with d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8 not found: ID does not exist" containerID="d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.548165 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8"} err="failed to get container status \"d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8\": rpc error: code = NotFound desc = could not find container \"d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8\": container with ID starting with d3f337931d2af352a55b5478429c52d366a4653a8af34fc9783de93981618ca8 not found: ID does not exist" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.548234 4822 scope.go:117] "RemoveContainer" containerID="98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35" Nov 24 15:22:38 crc kubenswrapper[4822]: E1124 15:22:38.548740 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35\": container with ID starting with 98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35 not found: ID does not exist" containerID="98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35" Nov 24 15:22:38 crc kubenswrapper[4822]: I1124 15:22:38.548763 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35"} err="failed to get container status \"98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35\": rpc error: code = NotFound desc = could not find container \"98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35\": container with ID starting with 98a23c4d9d2e54d41bb2818b39fdc403808a76cac2b0bfdb59c213175f17ac35 not found: ID does not exist" Nov 24 15:22:39 crc kubenswrapper[4822]: I1124 15:22:39.726105 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" path="/var/lib/kubelet/pods/02c1c1e6-79ea-48e3-bd98-c46f7538d205/volumes" Nov 24 15:22:46 crc kubenswrapper[4822]: I1124 15:22:46.525516 4822 generic.go:334] "Generic (PLEG): container finished" podID="b810e4f5-1dc3-4713-8a8c-18d816697cef" containerID="9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c" exitCode=0 Nov 24 15:22:46 crc kubenswrapper[4822]: I1124 15:22:46.525724 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-94kzw/must-gather-l9b7n" event={"ID":"b810e4f5-1dc3-4713-8a8c-18d816697cef","Type":"ContainerDied","Data":"9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c"} Nov 24 15:22:46 crc kubenswrapper[4822]: I1124 15:22:46.527261 4822 scope.go:117] "RemoveContainer" containerID="9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c" Nov 24 15:22:47 crc kubenswrapper[4822]: I1124 15:22:47.139485 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-94kzw_must-gather-l9b7n_b810e4f5-1dc3-4713-8a8c-18d816697cef/gather/0.log" Nov 24 15:22:49 crc kubenswrapper[4822]: I1124 15:22:49.705739 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:22:49 crc kubenswrapper[4822]: E1124 15:22:49.707027 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:22:54 crc kubenswrapper[4822]: I1124 15:22:54.596129 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-94kzw/must-gather-l9b7n"] Nov 24 15:22:54 crc kubenswrapper[4822]: I1124 15:22:54.596951 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-94kzw/must-gather-l9b7n" podUID="b810e4f5-1dc3-4713-8a8c-18d816697cef" containerName="copy" containerID="cri-o://0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6" gracePeriod=2 Nov 24 15:22:54 crc kubenswrapper[4822]: I1124 15:22:54.615484 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-94kzw/must-gather-l9b7n"] Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.092816 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-94kzw_must-gather-l9b7n_b810e4f5-1dc3-4713-8a8c-18d816697cef/copy/0.log" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.093637 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.260125 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7wms\" (UniqueName: \"kubernetes.io/projected/b810e4f5-1dc3-4713-8a8c-18d816697cef-kube-api-access-n7wms\") pod \"b810e4f5-1dc3-4713-8a8c-18d816697cef\" (UID: \"b810e4f5-1dc3-4713-8a8c-18d816697cef\") " Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.260387 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b810e4f5-1dc3-4713-8a8c-18d816697cef-must-gather-output\") pod \"b810e4f5-1dc3-4713-8a8c-18d816697cef\" (UID: \"b810e4f5-1dc3-4713-8a8c-18d816697cef\") " Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.267426 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b810e4f5-1dc3-4713-8a8c-18d816697cef-kube-api-access-n7wms" (OuterVolumeSpecName: "kube-api-access-n7wms") pod "b810e4f5-1dc3-4713-8a8c-18d816697cef" (UID: "b810e4f5-1dc3-4713-8a8c-18d816697cef"). InnerVolumeSpecName "kube-api-access-n7wms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.362992 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7wms\" (UniqueName: \"kubernetes.io/projected/b810e4f5-1dc3-4713-8a8c-18d816697cef-kube-api-access-n7wms\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.412828 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b810e4f5-1dc3-4713-8a8c-18d816697cef-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b810e4f5-1dc3-4713-8a8c-18d816697cef" (UID: "b810e4f5-1dc3-4713-8a8c-18d816697cef"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.464964 4822 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b810e4f5-1dc3-4713-8a8c-18d816697cef-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.646589 4822 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-94kzw_must-gather-l9b7n_b810e4f5-1dc3-4713-8a8c-18d816697cef/copy/0.log" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.648053 4822 generic.go:334] "Generic (PLEG): container finished" podID="b810e4f5-1dc3-4713-8a8c-18d816697cef" containerID="0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6" exitCode=143 Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.648150 4822 scope.go:117] "RemoveContainer" containerID="0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.648109 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-94kzw/must-gather-l9b7n" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.680582 4822 scope.go:117] "RemoveContainer" containerID="9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.726593 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b810e4f5-1dc3-4713-8a8c-18d816697cef" path="/var/lib/kubelet/pods/b810e4f5-1dc3-4713-8a8c-18d816697cef/volumes" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.777483 4822 scope.go:117] "RemoveContainer" containerID="0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6" Nov 24 15:22:55 crc kubenswrapper[4822]: E1124 15:22:55.777948 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6\": container with ID starting with 0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6 not found: ID does not exist" containerID="0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.777986 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6"} err="failed to get container status \"0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6\": rpc error: code = NotFound desc = could not find container \"0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6\": container with ID starting with 0e0552ebfd8493cd5a8d60545b8dd477688d58790a5d7248efd3ee1c842968b6 not found: ID does not exist" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.778012 4822 scope.go:117] "RemoveContainer" containerID="9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c" Nov 24 15:22:55 crc kubenswrapper[4822]: E1124 15:22:55.778400 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c\": container with ID starting with 9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c not found: ID does not exist" containerID="9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c" Nov 24 15:22:55 crc kubenswrapper[4822]: I1124 15:22:55.778441 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c"} err="failed to get container status \"9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c\": rpc error: code = NotFound desc = could not find container \"9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c\": container with ID starting with 9f84943400fba5064006394c162db74b204e671eaff103806cc2e66f2350780c not found: ID does not exist" Nov 24 15:23:02 crc kubenswrapper[4822]: I1124 15:23:02.705519 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:23:02 crc kubenswrapper[4822]: E1124 15:23:02.706352 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:23:13 crc kubenswrapper[4822]: I1124 15:23:13.712396 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:23:13 crc kubenswrapper[4822]: E1124 15:23:13.713329 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:23:27 crc kubenswrapper[4822]: I1124 15:23:27.705491 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:23:27 crc kubenswrapper[4822]: E1124 15:23:27.706254 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:23:41 crc kubenswrapper[4822]: I1124 15:23:41.705254 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:23:41 crc kubenswrapper[4822]: E1124 15:23:41.706726 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:23:56 crc kubenswrapper[4822]: I1124 15:23:56.704777 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:23:56 crc kubenswrapper[4822]: E1124 15:23:56.705522 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:24:10 crc kubenswrapper[4822]: I1124 15:24:10.705833 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:24:10 crc kubenswrapper[4822]: E1124 15:24:10.708191 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:24:24 crc kubenswrapper[4822]: I1124 15:24:24.704608 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:24:24 crc kubenswrapper[4822]: E1124 15:24:24.705688 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:24:36 crc kubenswrapper[4822]: I1124 15:24:36.705931 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:24:36 crc kubenswrapper[4822]: E1124 15:24:36.707140 4822 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nst99_openshift-machine-config-operator(9cb8d340-5782-4aa4-b81e-e1b64e6b5207)\"" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" Nov 24 15:24:49 crc kubenswrapper[4822]: I1124 15:24:49.704509 4822 scope.go:117] "RemoveContainer" containerID="e676bc8a56698977f4027b5049f2d7a5383bbb9dc199cc83fffb38fee12450e5" Nov 24 15:24:50 crc kubenswrapper[4822]: I1124 15:24:50.050569 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nst99" event={"ID":"9cb8d340-5782-4aa4-b81e-e1b64e6b5207","Type":"ContainerStarted","Data":"d4eee9e3f1ce82b39873ca1230f393cf9a31be34bbd1957be6be037cada936b2"} Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.321408 4822 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jg5jk"] Nov 24 15:25:13 crc kubenswrapper[4822]: E1124 15:25:13.322702 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b810e4f5-1dc3-4713-8a8c-18d816697cef" containerName="copy" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.322725 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b810e4f5-1dc3-4713-8a8c-18d816697cef" containerName="copy" Nov 24 15:25:13 crc kubenswrapper[4822]: E1124 15:25:13.322750 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerName="extract-content" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.322760 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerName="extract-content" Nov 24 15:25:13 crc kubenswrapper[4822]: E1124 15:25:13.322781 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerName="extract-utilities" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.322793 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerName="extract-utilities" Nov 24 15:25:13 crc kubenswrapper[4822]: E1124 15:25:13.322816 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerName="registry-server" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.322827 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerName="registry-server" Nov 24 15:25:13 crc kubenswrapper[4822]: E1124 15:25:13.322841 4822 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b810e4f5-1dc3-4713-8a8c-18d816697cef" containerName="gather" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.322850 4822 state_mem.go:107] "Deleted CPUSet assignment" podUID="b810e4f5-1dc3-4713-8a8c-18d816697cef" containerName="gather" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.323175 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="b810e4f5-1dc3-4713-8a8c-18d816697cef" containerName="gather" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.323226 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="02c1c1e6-79ea-48e3-bd98-c46f7538d205" containerName="registry-server" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.323264 4822 memory_manager.go:354] "RemoveStaleState removing state" podUID="b810e4f5-1dc3-4713-8a8c-18d816697cef" containerName="copy" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.327186 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.351614 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jg5jk"] Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.420598 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-utilities\") pod \"community-operators-jg5jk\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.420653 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znccz\" (UniqueName: \"kubernetes.io/projected/d09a1e09-312b-49e4-9962-ff067f66e868-kube-api-access-znccz\") pod \"community-operators-jg5jk\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.420955 4822 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-catalog-content\") pod \"community-operators-jg5jk\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.523366 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znccz\" (UniqueName: \"kubernetes.io/projected/d09a1e09-312b-49e4-9962-ff067f66e868-kube-api-access-znccz\") pod \"community-operators-jg5jk\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.523693 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-catalog-content\") pod \"community-operators-jg5jk\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.523853 4822 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-utilities\") pod \"community-operators-jg5jk\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.524196 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-catalog-content\") pod \"community-operators-jg5jk\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.524268 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-utilities\") pod \"community-operators-jg5jk\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.545862 4822 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znccz\" (UniqueName: \"kubernetes.io/projected/d09a1e09-312b-49e4-9962-ff067f66e868-kube-api-access-znccz\") pod \"community-operators-jg5jk\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:13 crc kubenswrapper[4822]: I1124 15:25:13.660493 4822 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:14 crc kubenswrapper[4822]: I1124 15:25:14.208007 4822 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jg5jk"] Nov 24 15:25:14 crc kubenswrapper[4822]: I1124 15:25:14.394816 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jg5jk" event={"ID":"d09a1e09-312b-49e4-9962-ff067f66e868","Type":"ContainerStarted","Data":"54652c5c56174084a51e80717d84dc7410dbc60325c4034beda1ff7684698cd6"} Nov 24 15:25:15 crc kubenswrapper[4822]: I1124 15:25:15.406542 4822 generic.go:334] "Generic (PLEG): container finished" podID="d09a1e09-312b-49e4-9962-ff067f66e868" containerID="168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230" exitCode=0 Nov 24 15:25:15 crc kubenswrapper[4822]: I1124 15:25:15.406729 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jg5jk" event={"ID":"d09a1e09-312b-49e4-9962-ff067f66e868","Type":"ContainerDied","Data":"168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230"} Nov 24 15:25:16 crc kubenswrapper[4822]: I1124 15:25:16.420667 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jg5jk" event={"ID":"d09a1e09-312b-49e4-9962-ff067f66e868","Type":"ContainerStarted","Data":"3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550"} Nov 24 15:25:17 crc kubenswrapper[4822]: I1124 15:25:17.433671 4822 generic.go:334] "Generic (PLEG): container finished" podID="d09a1e09-312b-49e4-9962-ff067f66e868" containerID="3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550" exitCode=0 Nov 24 15:25:17 crc kubenswrapper[4822]: I1124 15:25:17.433734 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jg5jk" event={"ID":"d09a1e09-312b-49e4-9962-ff067f66e868","Type":"ContainerDied","Data":"3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550"} Nov 24 15:25:18 crc kubenswrapper[4822]: I1124 15:25:18.446880 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jg5jk" event={"ID":"d09a1e09-312b-49e4-9962-ff067f66e868","Type":"ContainerStarted","Data":"3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70"} Nov 24 15:25:18 crc kubenswrapper[4822]: I1124 15:25:18.473744 4822 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jg5jk" podStartSLOduration=2.995719498 podStartE2EDuration="5.473724669s" podCreationTimestamp="2025-11-24 15:25:13 +0000 UTC" firstStartedPulling="2025-11-24 15:25:15.408858783 +0000 UTC m=+3952.525499260" lastFinishedPulling="2025-11-24 15:25:17.886863954 +0000 UTC m=+3955.003504431" observedRunningTime="2025-11-24 15:25:18.470240929 +0000 UTC m=+3955.586881416" watchObservedRunningTime="2025-11-24 15:25:18.473724669 +0000 UTC m=+3955.590365156" Nov 24 15:25:23 crc kubenswrapper[4822]: I1124 15:25:23.661120 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:23 crc kubenswrapper[4822]: I1124 15:25:23.661561 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:23 crc kubenswrapper[4822]: I1124 15:25:23.738664 4822 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:24 crc kubenswrapper[4822]: I1124 15:25:24.596317 4822 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:24 crc kubenswrapper[4822]: I1124 15:25:24.658070 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jg5jk"] Nov 24 15:25:26 crc kubenswrapper[4822]: I1124 15:25:26.543780 4822 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jg5jk" podUID="d09a1e09-312b-49e4-9962-ff067f66e868" containerName="registry-server" containerID="cri-o://3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70" gracePeriod=2 Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.197241 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.246363 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-utilities\") pod \"d09a1e09-312b-49e4-9962-ff067f66e868\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.246500 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znccz\" (UniqueName: \"kubernetes.io/projected/d09a1e09-312b-49e4-9962-ff067f66e868-kube-api-access-znccz\") pod \"d09a1e09-312b-49e4-9962-ff067f66e868\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.246744 4822 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-catalog-content\") pod \"d09a1e09-312b-49e4-9962-ff067f66e868\" (UID: \"d09a1e09-312b-49e4-9962-ff067f66e868\") " Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.248011 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-utilities" (OuterVolumeSpecName: "utilities") pod "d09a1e09-312b-49e4-9962-ff067f66e868" (UID: "d09a1e09-312b-49e4-9962-ff067f66e868"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.254447 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d09a1e09-312b-49e4-9962-ff067f66e868-kube-api-access-znccz" (OuterVolumeSpecName: "kube-api-access-znccz") pod "d09a1e09-312b-49e4-9962-ff067f66e868" (UID: "d09a1e09-312b-49e4-9962-ff067f66e868"). InnerVolumeSpecName "kube-api-access-znccz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.308049 4822 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d09a1e09-312b-49e4-9962-ff067f66e868" (UID: "d09a1e09-312b-49e4-9962-ff067f66e868"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.348856 4822 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.348909 4822 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znccz\" (UniqueName: \"kubernetes.io/projected/d09a1e09-312b-49e4-9962-ff067f66e868-kube-api-access-znccz\") on node \"crc\" DevicePath \"\"" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.348923 4822 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d09a1e09-312b-49e4-9962-ff067f66e868-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.553588 4822 generic.go:334] "Generic (PLEG): container finished" podID="d09a1e09-312b-49e4-9962-ff067f66e868" containerID="3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70" exitCode=0 Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.553791 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jg5jk" event={"ID":"d09a1e09-312b-49e4-9962-ff067f66e868","Type":"ContainerDied","Data":"3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70"} Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.554596 4822 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jg5jk" event={"ID":"d09a1e09-312b-49e4-9962-ff067f66e868","Type":"ContainerDied","Data":"54652c5c56174084a51e80717d84dc7410dbc60325c4034beda1ff7684698cd6"} Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.553900 4822 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jg5jk" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.554665 4822 scope.go:117] "RemoveContainer" containerID="3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.596672 4822 scope.go:117] "RemoveContainer" containerID="3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.624997 4822 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jg5jk"] Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.626581 4822 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jg5jk"] Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.627989 4822 scope.go:117] "RemoveContainer" containerID="168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.675511 4822 scope.go:117] "RemoveContainer" containerID="3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70" Nov 24 15:25:27 crc kubenswrapper[4822]: E1124 15:25:27.675965 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70\": container with ID starting with 3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70 not found: ID does not exist" containerID="3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.676041 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70"} err="failed to get container status \"3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70\": rpc error: code = NotFound desc = could not find container \"3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70\": container with ID starting with 3556db934b5cf06f99b65e22e1c4a2f4018109f51830177e51c8235f14390e70 not found: ID does not exist" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.676074 4822 scope.go:117] "RemoveContainer" containerID="3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550" Nov 24 15:25:27 crc kubenswrapper[4822]: E1124 15:25:27.676369 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550\": container with ID starting with 3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550 not found: ID does not exist" containerID="3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.676406 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550"} err="failed to get container status \"3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550\": rpc error: code = NotFound desc = could not find container \"3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550\": container with ID starting with 3bd73732118b2f3aa3dca5b84bb151fa8353522a0abbd9b861e91f45c6b0b550 not found: ID does not exist" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.676427 4822 scope.go:117] "RemoveContainer" containerID="168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230" Nov 24 15:25:27 crc kubenswrapper[4822]: E1124 15:25:27.676806 4822 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230\": container with ID starting with 168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230 not found: ID does not exist" containerID="168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.676838 4822 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230"} err="failed to get container status \"168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230\": rpc error: code = NotFound desc = could not find container \"168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230\": container with ID starting with 168a2328baad68323db7935fdbf49b395927d22f9e856745e4132ff4cb468230 not found: ID does not exist" Nov 24 15:25:27 crc kubenswrapper[4822]: I1124 15:25:27.715420 4822 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d09a1e09-312b-49e4-9962-ff067f66e868" path="/var/lib/kubelet/pods/d09a1e09-312b-49e4-9962-ff067f66e868/volumes" Nov 24 15:26:47 crc kubenswrapper[4822]: I1124 15:26:47.116614 4822 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-ctks4 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 15:26:47 crc kubenswrapper[4822]: I1124 15:26:47.117198 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" podUID="90001d98-f4e4-42f3-8f78-453f244a6286" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 15:26:47 crc kubenswrapper[4822]: I1124 15:26:47.121799 4822 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-ctks4 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 15:26:47 crc kubenswrapper[4822]: I1124 15:26:47.121968 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ctks4" podUID="90001d98-f4e4-42f3-8f78-453f244a6286" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 15:26:47 crc kubenswrapper[4822]: I1124 15:26:47.339357 4822 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ftv2h container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 15:26:47 crc kubenswrapper[4822]: I1124 15:26:47.339680 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" podUID="81e0e571-ef24-47df-a054-54226fcfe214" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 15:26:47 crc kubenswrapper[4822]: I1124 15:26:47.362490 4822 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ftv2h container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 15:26:47 crc kubenswrapper[4822]: I1124 15:26:47.363269 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ftv2h" podUID="81e0e571-ef24-47df-a054-54226fcfe214" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 15:26:47 crc kubenswrapper[4822]: I1124 15:26:47.464795 4822 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-compactor-0" podUID="ac81e91d-a816-4a86-98b6-3e43f36c7b6f" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.128:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 15:27:11 crc kubenswrapper[4822]: I1124 15:27:11.577764 4822 patch_prober.go:28] interesting pod/machine-config-daemon-nst99 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:27:11 crc kubenswrapper[4822]: I1124 15:27:11.579193 4822 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nst99" podUID="9cb8d340-5782-4aa4-b81e-e1b64e6b5207" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111074551024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111074552017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111064231016477 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111064231015447 5ustar corecore